2026-03-07T07:50:15.032 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-07T07:50:15.038 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-07T07:50:15.058 INFO:teuthology.run:Config: archive_path: /archive/irq0-2026-03-07_00:06:21-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/502 branch: cobaltcore-storage-v19.2.3-fasttrack-5 description: orch:cephadm:smoke-roleless/{0-distro/ubuntu_22.04 1-start 2-services/rgw-ingress 3-final} email: null first_in_suite: false flavor: default job_id: '502' ktype: distro last_in_suite: false machine_type: vps name: irq0-2026-03-07_00:06:21-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps no_nested_subset: false openstack: - volumes: count: 4 size: 10 os_type: ubuntu os_version: '22.04' overrides: admin_socket: branch: cobaltcore-storage-v19.2.3-fasttrack-5 ansible.cephlab: branch: main repo: https://github.com/kshtsk/ceph-cm-ansible.git skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: Europe/Berlin ceph: conf: mgr: debug mgr: 20 debug ms: 1 mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 osd shutdown pgref assert: true flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - CEPHADM_DAEMON_PLACE_FAIL - CEPHADM_FAILED_DAEMON log-only-match: - CEPHADM_ sha1: 340d3c24fc6ae7529322dc7ccee6c6cb2589da0a ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} cephadm: cephadm_binary_url: https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm containers: image: harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 install: ceph: flavor: default sha1: 340d3c24fc6ae7529322dc7ccee6c6cb2589da0a extra_system_packages: deb: - python3-xmltodict - s3cmd rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - s3cmd repos: - name: ceph-source priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-39-g340d3c24fc6/el9.clyso/SRPMS - name: ceph-noarch priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-39-g340d3c24fc6/el9.clyso/noarch - name: ceph priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph/rpm-19.2.3-39-g340d3c24fc6/el9.clyso/x86_64 workunit: branch: tt-fasttrack-5-kafka sha1: e5dd18616ec0644f6e49fdd43c5c566bbe437a63 owner: irq0 priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - client.0 - - host.b - client.1 seed: 2469 sha1: 340d3c24fc6ae7529322dc7ccee6c6cb2589da0a sleep_before_teardown: 0 suite: orch:cephadm:smoke-roleless suite_branch: tt-fasttrack-5-kafka suite_path: /home/teuthos/src/github.com_kshtsk_ceph_e5dd18616ec0644f6e49fdd43c5c566bbe437a63/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: e5dd18616ec0644f6e49fdd43c5c566bbe437a63 targets: vm07.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLs9S6Vc1yvjO5ULo8cAPoW6v1uLpfhOpO4pmdbcX77+oevkfvURmOAKPYmbcQxNzWijHwnBU+9wo7qFs8Nk4pM= vm09.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBGRte6l2ocXJs/QsLJd2c5xEikXbnTZBf/TL+fdFh6NDbHrFQBfwATws7BMTglR7hAMJfT3q1zE+LXhppWDcTkM= tasks: - cephadm: roleless: true - cephadm.shell: host.a: - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - vip: null - cephadm.shell: host.a: - ceph orch device ls --refresh - cephadm.apply: specs: - placement: count: 4 host_pattern: '*' service_id: foo service_type: rgw spec: rgw_frontend_port: 8000 - placement: count: 2 service_id: rgw.foo service_type: ingress spec: backend_service: rgw.foo frontend_port: 9000 monitor_port: 9001 virtual_ip: '{{VIP0}}/{{VIPPREFIXLEN}}' - cephadm.wait_for_service: service: rgw.foo - cephadm.wait_for_service: service: ingress.rgw.foo - cephadm.shell: host.a: - "echo \"Check while healthy...\"\ncurl http://{{VIP0}}:9000/\n\n# stop each\ \ rgw in turn\necho \"Check with each rgw stopped in turn...\"\nfor rgw in `ceph\ \ orch ps | grep ^rgw.foo. | awk '{print $1}'`; do\n ceph orch daemon stop\ \ $rgw\n timeout 300 bash -c \"while ! ceph orch ps | grep $rgw | grep stopped;\ \ do echo 'Waiting for $rgw to stop'; ceph orch ps --daemon-type rgw; ceph health\ \ detail; sleep 5 ; done\"\n timeout 300 bash -c \"while ! curl http://{{VIP0}}:9000/\ \ ; do echo 'Waiting for http://{{VIP0}}:9000/ to be available'; sleep 1 ; done\"\ \n ceph orch daemon start $rgw\n timeout 300 bash -c \"while ! ceph orch ps\ \ | grep $rgw | grep running; do echo 'Waiting for $rgw to start'; ceph orch\ \ ps --daemon-type rgw; ceph health detail; sleep 5 ; done\"\ndone\n\n# stop\ \ each haproxy in turn\necho \"Check with each haproxy down in turn...\"\nfor\ \ haproxy in `ceph orch ps | grep ^haproxy.rgw.foo. | awk '{print $1}'`; do\n\ \ ceph orch daemon stop $haproxy\n timeout 300 bash -c \"while ! ceph orch\ \ ps | grep $haproxy | grep stopped; do echo 'Waiting for $haproxy to stop';\ \ ceph orch ps --daemon-type haproxy; ceph health detail; sleep 5 ; done\"\n\ \ timeout 300 bash -c \"while ! curl http://{{VIP0}}:9000/ ; do echo 'Waiting\ \ for http://{{VIP0}}:9000/ to be available'; sleep 1 ; done\"\n ceph orch\ \ daemon start $haproxy\n timeout 300 bash -c \"while ! ceph orch ps | grep\ \ $haproxy | grep running; do echo 'Waiting for $haproxy to start'; ceph orch\ \ ps --daemon-type haproxy; ceph health detail; sleep 5 ; done\"\ndone\n\ntimeout\ \ 300 bash -c \"while ! curl http://{{VIP0}}:9000/ ; do echo 'Waiting for http://{{VIP0}}:9000/\ \ to be available'; sleep 1 ; done\"\n" - cephadm.shell: host.a: - stat -c '%u %g' /var/log/ceph | grep '167 167' - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - ceph orch ls | grep '^osd.all-available-devices ' teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-07_00:06:21 tube: vps user: irq0 verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.43333 2026-03-07T07:50:15.058 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_e5dd18616ec0644f6e49fdd43c5c566bbe437a63/qa; will attempt to use it 2026-03-07T07:50:15.059 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_e5dd18616ec0644f6e49fdd43c5c566bbe437a63/qa/tasks 2026-03-07T07:50:15.059 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-07T07:50:15.060 INFO:teuthology.task.internal:Saving configuration 2026-03-07T07:50:15.067 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-07T07:50:15.068 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-07T07:50:15.074 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm07.local', 'description': '/archive/irq0-2026-03-07_00:06:21-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/502', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'ubuntu', 'os_version': '22.04', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-07 06:49:06.245180', 'locked_by': 'irq0', 'mac_address': '52:55:00:00:00:07', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLs9S6Vc1yvjO5ULo8cAPoW6v1uLpfhOpO4pmdbcX77+oevkfvURmOAKPYmbcQxNzWijHwnBU+9wo7qFs8Nk4pM='} 2026-03-07T07:50:15.079 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm09.local', 'description': '/archive/irq0-2026-03-07_00:06:21-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/502', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'ubuntu', 'os_version': '22.04', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-07 06:49:06.245742', 'locked_by': 'irq0', 'mac_address': '52:55:00:00:00:09', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBGRte6l2ocXJs/QsLJd2c5xEikXbnTZBf/TL+fdFh6NDbHrFQBfwATws7BMTglR7hAMJfT3q1zE+LXhppWDcTkM='} 2026-03-07T07:50:15.079 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-07T07:50:15.080 INFO:teuthology.task.internal:roles: ubuntu@vm07.local - ['host.a', 'client.0'] 2026-03-07T07:50:15.080 INFO:teuthology.task.internal:roles: ubuntu@vm09.local - ['host.b', 'client.1'] 2026-03-07T07:50:15.080 INFO:teuthology.run_tasks:Running task console_log... 2026-03-07T07:50:15.086 DEBUG:teuthology.task.console_log:vm07 does not support IPMI; excluding 2026-03-07T07:50:15.090 DEBUG:teuthology.task.console_log:vm09 does not support IPMI; excluding 2026-03-07T07:50:15.090 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7faeb5ff0820>, signals=[15]) 2026-03-07T07:50:15.091 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-07T07:50:15.091 INFO:teuthology.task.internal:Opening connections... 2026-03-07T07:50:15.091 DEBUG:teuthology.task.internal:connecting to ubuntu@vm07.local 2026-03-07T07:50:15.092 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm07.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-07T07:50:15.150 DEBUG:teuthology.task.internal:connecting to ubuntu@vm09.local 2026-03-07T07:50:15.151 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm09.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-07T07:50:15.207 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-07T07:50:15.208 DEBUG:teuthology.orchestra.run.vm07:> uname -m 2026-03-07T07:50:15.212 INFO:teuthology.orchestra.run.vm07.stdout:x86_64 2026-03-07T07:50:15.212 DEBUG:teuthology.orchestra.run.vm07:> cat /etc/os-release 2026-03-07T07:50:15.255 INFO:teuthology.orchestra.run.vm07.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-07T07:50:15.255 INFO:teuthology.orchestra.run.vm07.stdout:NAME="Ubuntu" 2026-03-07T07:50:15.255 INFO:teuthology.orchestra.run.vm07.stdout:VERSION_ID="22.04" 2026-03-07T07:50:15.255 INFO:teuthology.orchestra.run.vm07.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-07T07:50:15.255 INFO:teuthology.orchestra.run.vm07.stdout:VERSION_CODENAME=jammy 2026-03-07T07:50:15.255 INFO:teuthology.orchestra.run.vm07.stdout:ID=ubuntu 2026-03-07T07:50:15.255 INFO:teuthology.orchestra.run.vm07.stdout:ID_LIKE=debian 2026-03-07T07:50:15.255 INFO:teuthology.orchestra.run.vm07.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-07T07:50:15.256 INFO:teuthology.orchestra.run.vm07.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-07T07:50:15.256 INFO:teuthology.orchestra.run.vm07.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-07T07:50:15.256 INFO:teuthology.orchestra.run.vm07.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-07T07:50:15.256 INFO:teuthology.orchestra.run.vm07.stdout:UBUNTU_CODENAME=jammy 2026-03-07T07:50:15.256 INFO:teuthology.lock.ops:Updating vm07.local on lock server 2026-03-07T07:50:15.261 DEBUG:teuthology.orchestra.run.vm09:> uname -m 2026-03-07T07:50:15.289 INFO:teuthology.orchestra.run.vm09.stdout:x86_64 2026-03-07T07:50:15.290 DEBUG:teuthology.orchestra.run.vm09:> cat /etc/os-release 2026-03-07T07:50:15.334 INFO:teuthology.orchestra.run.vm09.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-07T07:50:15.335 INFO:teuthology.orchestra.run.vm09.stdout:NAME="Ubuntu" 2026-03-07T07:50:15.335 INFO:teuthology.orchestra.run.vm09.stdout:VERSION_ID="22.04" 2026-03-07T07:50:15.335 INFO:teuthology.orchestra.run.vm09.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-07T07:50:15.335 INFO:teuthology.orchestra.run.vm09.stdout:VERSION_CODENAME=jammy 2026-03-07T07:50:15.335 INFO:teuthology.orchestra.run.vm09.stdout:ID=ubuntu 2026-03-07T07:50:15.335 INFO:teuthology.orchestra.run.vm09.stdout:ID_LIKE=debian 2026-03-07T07:50:15.335 INFO:teuthology.orchestra.run.vm09.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-07T07:50:15.335 INFO:teuthology.orchestra.run.vm09.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-07T07:50:15.335 INFO:teuthology.orchestra.run.vm09.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-07T07:50:15.335 INFO:teuthology.orchestra.run.vm09.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-07T07:50:15.335 INFO:teuthology.orchestra.run.vm09.stdout:UBUNTU_CODENAME=jammy 2026-03-07T07:50:15.335 INFO:teuthology.lock.ops:Updating vm09.local on lock server 2026-03-07T07:50:15.339 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-07T07:50:15.341 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-07T07:50:15.341 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-07T07:50:15.341 DEBUG:teuthology.orchestra.run.vm07:> test '!' -e /home/ubuntu/cephtest 2026-03-07T07:50:15.342 DEBUG:teuthology.orchestra.run.vm09:> test '!' -e /home/ubuntu/cephtest 2026-03-07T07:50:15.378 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-07T07:50:15.379 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-07T07:50:15.379 DEBUG:teuthology.orchestra.run.vm07:> test -z $(ls -A /var/lib/ceph) 2026-03-07T07:50:15.385 DEBUG:teuthology.orchestra.run.vm09:> test -z $(ls -A /var/lib/ceph) 2026-03-07T07:50:15.387 INFO:teuthology.orchestra.run.vm07.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-07T07:50:15.423 INFO:teuthology.orchestra.run.vm09.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-07T07:50:15.423 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-07T07:50:15.432 DEBUG:teuthology.orchestra.run.vm07:> test -e /ceph-qa-ready 2026-03-07T07:50:15.434 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-07T07:50:15.673 DEBUG:teuthology.orchestra.run.vm09:> test -e /ceph-qa-ready 2026-03-07T07:50:15.676 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-07T07:50:15.985 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-07T07:50:15.986 INFO:teuthology.task.internal:Creating test directory... 2026-03-07T07:50:15.986 DEBUG:teuthology.orchestra.run.vm07:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-07T07:50:15.987 DEBUG:teuthology.orchestra.run.vm09:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-07T07:50:15.990 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-07T07:50:15.991 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-07T07:50:15.992 INFO:teuthology.task.internal:Creating archive directory... 2026-03-07T07:50:15.992 DEBUG:teuthology.orchestra.run.vm07:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-07T07:50:16.033 DEBUG:teuthology.orchestra.run.vm09:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-07T07:50:16.037 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-07T07:50:16.038 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-07T07:50:16.038 DEBUG:teuthology.orchestra.run.vm07:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-07T07:50:16.078 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-07T07:50:16.078 DEBUG:teuthology.orchestra.run.vm09:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-07T07:50:16.082 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-07T07:50:16.082 DEBUG:teuthology.orchestra.run.vm07:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-07T07:50:16.121 DEBUG:teuthology.orchestra.run.vm09:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-07T07:50:16.128 INFO:teuthology.orchestra.run.vm07.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-07T07:50:16.131 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-07T07:50:16.132 INFO:teuthology.orchestra.run.vm07.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-07T07:50:16.136 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-07T07:50:16.137 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-07T07:50:16.138 INFO:teuthology.task.internal:Configuring sudo... 2026-03-07T07:50:16.138 DEBUG:teuthology.orchestra.run.vm07:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-07T07:50:16.177 DEBUG:teuthology.orchestra.run.vm09:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-07T07:50:16.187 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-07T07:50:16.189 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-07T07:50:16.189 DEBUG:teuthology.orchestra.run.vm07:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-07T07:50:16.225 DEBUG:teuthology.orchestra.run.vm09:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-07T07:50:16.230 DEBUG:teuthology.orchestra.run.vm07:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-07T07:50:16.271 DEBUG:teuthology.orchestra.run.vm07:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-07T07:50:16.315 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-07T07:50:16.315 DEBUG:teuthology.orchestra.run.vm07:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-07T07:50:16.364 DEBUG:teuthology.orchestra.run.vm09:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-07T07:50:16.367 DEBUG:teuthology.orchestra.run.vm09:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-07T07:50:16.410 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-07T07:50:16.410 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-07T07:50:16.459 DEBUG:teuthology.orchestra.run.vm07:> sudo service rsyslog restart 2026-03-07T07:50:16.460 DEBUG:teuthology.orchestra.run.vm09:> sudo service rsyslog restart 2026-03-07T07:50:16.515 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-07T07:50:16.516 INFO:teuthology.task.internal:Starting timer... 2026-03-07T07:50:16.516 INFO:teuthology.run_tasks:Running task pcp... 2026-03-07T07:50:16.520 INFO:teuthology.run_tasks:Running task selinux... 2026-03-07T07:50:16.522 INFO:teuthology.task.selinux:Excluding vm07: VMs are not yet supported 2026-03-07T07:50:16.522 INFO:teuthology.task.selinux:Excluding vm09: VMs are not yet supported 2026-03-07T07:50:16.522 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-07T07:50:16.522 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-07T07:50:16.522 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-07T07:50:16.522 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-07T07:50:16.523 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'repo': 'https://github.com/kshtsk/ceph-cm-ansible.git', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'Europe/Berlin'}} 2026-03-07T07:50:16.524 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/kshtsk/ceph-cm-ansible.git 2026-03-07T07:50:16.525 INFO:teuthology.repo_utils:Fetching github.com_kshtsk_ceph-cm-ansible_main from origin 2026-03-07T07:50:17.098 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main to origin/main 2026-03-07T07:50:17.114 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-07T07:50:17.114 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "Europe/Berlin"}' -i /tmp/teuth_ansible_inventory_vdp1sa2 --limit vm07.local,vm09.local /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-07T07:52:12.624 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm07.local'), Remote(name='ubuntu@vm09.local')] 2026-03-07T07:52:12.625 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm07.local' 2026-03-07T07:52:12.625 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm07.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-07T07:52:12.689 DEBUG:teuthology.orchestra.run.vm07:> true 2026-03-07T07:52:12.908 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm07.local' 2026-03-07T07:52:12.908 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm09.local' 2026-03-07T07:52:12.908 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm09.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-07T07:52:12.969 DEBUG:teuthology.orchestra.run.vm09:> true 2026-03-07T07:52:13.168 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm09.local' 2026-03-07T07:52:13.168 INFO:teuthology.run_tasks:Running task clock... 2026-03-07T07:52:13.171 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-07T07:52:13.171 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-07T07:52:13.171 DEBUG:teuthology.orchestra.run.vm07:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-07T07:52:13.172 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-07T07:52:13.172 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-07T07:52:13.187 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-07T07:52:13.188 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: Command line: ntpd -gq 2026-03-07T07:52:13.188 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: ---------------------------------------------------- 2026-03-07T07:52:13.188 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: ntp-4 is maintained by Network Time Foundation, 2026-03-07T07:52:13.188 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-07T07:52:13.188 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: corporation. Support and training for ntp-4 are 2026-03-07T07:52:13.188 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: available at https://www.nwtime.org/support 2026-03-07T07:52:13.188 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: ---------------------------------------------------- 2026-03-07T07:52:13.188 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: proto: precision = 0.029 usec (-25) 2026-03-07T07:52:13.188 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: basedate set to 2022-02-04 2026-03-07T07:52:13.188 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: gps base set to 2022-02-06 (week 2196) 2026-03-07T07:52:13.188 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-07T07:52:13.188 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-07T07:52:13.188 INFO:teuthology.orchestra.run.vm07.stderr: 7 Mar 07:52:13 ntpd[15656]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 70 days ago 2026-03-07T07:52:13.188 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: Listen and drop on 0 v6wildcard [::]:123 2026-03-07T07:52:13.188 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-07T07:52:13.189 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: Listen normally on 2 lo 127.0.0.1:123 2026-03-07T07:52:13.189 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: Listen normally on 3 ens3 192.168.123.107:123 2026-03-07T07:52:13.189 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: Listen normally on 4 lo [::1]:123 2026-03-07T07:52:13.189 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:7%2]:123 2026-03-07T07:52:13.189 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:13 ntpd[15656]: Listening on routing socket on fd #22 for interface updates 2026-03-07T07:52:13.225 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-07T07:52:13.225 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: Command line: ntpd -gq 2026-03-07T07:52:13.225 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: ---------------------------------------------------- 2026-03-07T07:52:13.225 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: ntp-4 is maintained by Network Time Foundation, 2026-03-07T07:52:13.225 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-07T07:52:13.225 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: corporation. Support and training for ntp-4 are 2026-03-07T07:52:13.225 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: available at https://www.nwtime.org/support 2026-03-07T07:52:13.225 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: ---------------------------------------------------- 2026-03-07T07:52:13.226 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: proto: precision = 0.029 usec (-25) 2026-03-07T07:52:13.226 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: basedate set to 2022-02-04 2026-03-07T07:52:13.226 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: gps base set to 2022-02-06 (week 2196) 2026-03-07T07:52:13.226 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-07T07:52:13.226 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-07T07:52:13.226 INFO:teuthology.orchestra.run.vm09.stderr: 7 Mar 07:52:13 ntpd[15637]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 70 days ago 2026-03-07T07:52:13.227 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: Listen and drop on 0 v6wildcard [::]:123 2026-03-07T07:52:13.227 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-07T07:52:13.227 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: Listen normally on 2 lo 127.0.0.1:123 2026-03-07T07:52:13.227 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: Listen normally on 3 ens3 192.168.123.109:123 2026-03-07T07:52:13.227 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: Listen normally on 4 lo [::1]:123 2026-03-07T07:52:13.228 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:9%2]:123 2026-03-07T07:52:13.228 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:13 ntpd[15637]: Listening on routing socket on fd #22 for interface updates 2026-03-07T07:52:14.188 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:14 ntpd[15656]: Soliciting pool server 212.132.97.26 2026-03-07T07:52:14.226 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:14 ntpd[15637]: Soliciting pool server 212.132.97.26 2026-03-07T07:52:15.187 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:15 ntpd[15656]: Soliciting pool server 85.215.189.120 2026-03-07T07:52:15.187 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:15 ntpd[15656]: Soliciting pool server 31.209.85.242 2026-03-07T07:52:15.225 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:15 ntpd[15637]: Soliciting pool server 85.215.189.120 2026-03-07T07:52:15.225 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:15 ntpd[15637]: Soliciting pool server 31.209.85.242 2026-03-07T07:52:16.187 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:16 ntpd[15656]: Soliciting pool server 144.76.59.37 2026-03-07T07:52:16.187 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:16 ntpd[15656]: Soliciting pool server 185.41.106.152 2026-03-07T07:52:16.187 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:16 ntpd[15656]: Soliciting pool server 176.9.44.212 2026-03-07T07:52:16.225 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:16 ntpd[15637]: Soliciting pool server 144.76.59.37 2026-03-07T07:52:16.225 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:16 ntpd[15637]: Soliciting pool server 185.41.106.152 2026-03-07T07:52:16.225 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:16 ntpd[15637]: Soliciting pool server 176.9.44.212 2026-03-07T07:52:17.186 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:17 ntpd[15656]: Soliciting pool server 130.61.89.107 2026-03-07T07:52:17.186 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:17 ntpd[15656]: Soliciting pool server 195.201.107.151 2026-03-07T07:52:17.186 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:17 ntpd[15656]: Soliciting pool server 185.11.138.90 2026-03-07T07:52:17.186 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:17 ntpd[15656]: Soliciting pool server 45.9.61.155 2026-03-07T07:52:17.224 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:17 ntpd[15637]: Soliciting pool server 130.61.89.107 2026-03-07T07:52:17.224 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:17 ntpd[15637]: Soliciting pool server 195.201.107.151 2026-03-07T07:52:17.224 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:17 ntpd[15637]: Soliciting pool server 45.9.61.155 2026-03-07T07:52:18.186 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:18 ntpd[15656]: Soliciting pool server 162.159.200.123 2026-03-07T07:52:18.186 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:18 ntpd[15656]: Soliciting pool server 139.162.152.20 2026-03-07T07:52:18.186 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:18 ntpd[15656]: Soliciting pool server 77.90.0.148 2026-03-07T07:52:18.186 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:18 ntpd[15656]: Soliciting pool server 185.125.190.56 2026-03-07T07:52:18.224 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:18 ntpd[15637]: Soliciting pool server 162.159.200.123 2026-03-07T07:52:18.224 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:18 ntpd[15637]: Soliciting pool server 139.162.152.20 2026-03-07T07:52:18.224 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:18 ntpd[15637]: Soliciting pool server 77.90.0.148 2026-03-07T07:52:18.224 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:18 ntpd[15637]: Soliciting pool server 185.125.190.56 2026-03-07T07:52:19.185 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:19 ntpd[15656]: Soliciting pool server 91.189.91.157 2026-03-07T07:52:19.185 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:19 ntpd[15656]: Soliciting pool server 116.203.96.227 2026-03-07T07:52:19.185 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:19 ntpd[15656]: Soliciting pool server 85.215.166.214 2026-03-07T07:52:19.223 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:19 ntpd[15637]: Soliciting pool server 91.189.91.157 2026-03-07T07:52:19.223 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:19 ntpd[15637]: Soliciting pool server 116.203.96.227 2026-03-07T07:52:19.223 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:19 ntpd[15637]: Soliciting pool server 85.215.166.214 2026-03-07T07:52:20.222 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:20 ntpd[15637]: Soliciting pool server 185.125.190.58 2026-03-07T07:52:20.223 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:20 ntpd[15637]: Soliciting pool server 62.108.36.235 2026-03-07T07:52:20.223 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:20 ntpd[15637]: Soliciting pool server 2003:a:87f:c37c::8 2026-03-07T07:52:21.213 INFO:teuthology.orchestra.run.vm07.stdout: 7 Mar 07:52:21 ntpd[15656]: ntpd: time slew +0.003516 s 2026-03-07T07:52:21.213 INFO:teuthology.orchestra.run.vm07.stdout:ntpd: time slew +0.003516s 2026-03-07T07:52:21.233 INFO:teuthology.orchestra.run.vm07.stdout: remote refid st t when poll reach delay offset jitter 2026-03-07T07:52:21.233 INFO:teuthology.orchestra.run.vm07.stdout:============================================================================== 2026-03-07T07:52:21.233 INFO:teuthology.orchestra.run.vm07.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T07:52:21.233 INFO:teuthology.orchestra.run.vm07.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T07:52:21.233 INFO:teuthology.orchestra.run.vm07.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T07:52:21.233 INFO:teuthology.orchestra.run.vm07.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T07:52:21.233 INFO:teuthology.orchestra.run.vm07.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T07:52:22.252 INFO:teuthology.orchestra.run.vm09.stdout: 7 Mar 07:52:22 ntpd[15637]: ntpd: time slew +0.021013 s 2026-03-07T07:52:22.252 INFO:teuthology.orchestra.run.vm09.stdout:ntpd: time slew +0.021013s 2026-03-07T07:52:22.271 INFO:teuthology.orchestra.run.vm09.stdout: remote refid st t when poll reach delay offset jitter 2026-03-07T07:52:22.271 INFO:teuthology.orchestra.run.vm09.stdout:============================================================================== 2026-03-07T07:52:22.271 INFO:teuthology.orchestra.run.vm09.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T07:52:22.271 INFO:teuthology.orchestra.run.vm09.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T07:52:22.271 INFO:teuthology.orchestra.run.vm09.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T07:52:22.271 INFO:teuthology.orchestra.run.vm09.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T07:52:22.271 INFO:teuthology.orchestra.run.vm09.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T07:52:22.271 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-07T07:52:22.319 INFO:tasks.cephadm:Config: {'roleless': True, 'conf': {'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000, 'osd shutdown pgref assert': True}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'CEPHADM_DAEMON_PLACE_FAIL', 'CEPHADM_FAILED_DAEMON'], 'log-only-match': ['CEPHADM_'], 'sha1': '340d3c24fc6ae7529322dc7ccee6c6cb2589da0a', 'cephadm_binary_url': 'https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm', 'containers': {'image': 'harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5'}} 2026-03-07T07:52:22.319 INFO:tasks.cephadm:Provided image contains tag or digest, using it as is 2026-03-07T07:52:22.319 INFO:tasks.cephadm:Cluster image is harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 2026-03-07T07:52:22.319 INFO:tasks.cephadm:Cluster fsid is 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:52:22.319 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-07T07:52:22.319 INFO:tasks.cephadm:No mon roles; fabricating mons 2026-03-07T07:52:22.319 INFO:tasks.cephadm:Monitor IPs: {'mon.vm07': '192.168.123.107', 'mon.vm09': '192.168.123.109'} 2026-03-07T07:52:22.319 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-07T07:52:22.319 DEBUG:teuthology.orchestra.run.vm07:> sudo hostname $(hostname -s) 2026-03-07T07:52:22.326 DEBUG:teuthology.orchestra.run.vm09:> sudo hostname $(hostname -s) 2026-03-07T07:52:22.333 INFO:tasks.cephadm:Downloading cephadm from url: https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm 2026-03-07T07:52:22.333 DEBUG:teuthology.orchestra.run.vm07:> curl --silent -L https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-07T07:52:23.430 INFO:teuthology.orchestra.run.vm07.stdout:-rw-rw-r-- 1 ubuntu ubuntu 787672 Mar 7 07:52 /home/ubuntu/cephtest/cephadm 2026-03-07T07:52:23.430 DEBUG:teuthology.orchestra.run.vm09:> curl --silent -L https://download.ceph.com/rpm-19.2.3/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-07T07:52:24.491 INFO:teuthology.orchestra.run.vm09.stdout:-rw-rw-r-- 1 ubuntu ubuntu 787672 Mar 7 07:52 /home/ubuntu/cephtest/cephadm 2026-03-07T07:52:24.491 DEBUG:teuthology.orchestra.run.vm07:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-07T07:52:24.495 DEBUG:teuthology.orchestra.run.vm09:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-07T07:52:24.501 INFO:tasks.cephadm:Pulling image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 on all hosts... 2026-03-07T07:52:24.501 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 pull 2026-03-07T07:52:24.537 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 pull 2026-03-07T07:52:24.776 INFO:teuthology.orchestra.run.vm07.stderr:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5... 2026-03-07T07:52:24.779 INFO:teuthology.orchestra.run.vm09.stderr:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5... 2026-03-07T07:52:49.074 INFO:teuthology.orchestra.run.vm09.stdout:{ 2026-03-07T07:52:49.074 INFO:teuthology.orchestra.run.vm09.stdout: "ceph_version": "ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable)", 2026-03-07T07:52:49.074 INFO:teuthology.orchestra.run.vm09.stdout: "image_id": "8bccc98d839aa18345ec1336292d0452ca331737e49f12524f635044dcabcfe1", 2026-03-07T07:52:49.074 INFO:teuthology.orchestra.run.vm09.stdout: "repo_digests": [ 2026-03-07T07:52:49.074 INFO:teuthology.orchestra.run.vm09.stdout: "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:ffa52c72fad7bdd2657408de9cf8d87fc2c72f716d1a00277ba13f7c12b404e0" 2026-03-07T07:52:49.074 INFO:teuthology.orchestra.run.vm09.stdout: ] 2026-03-07T07:52:49.074 INFO:teuthology.orchestra.run.vm09.stdout:} 2026-03-07T07:52:49.144 INFO:teuthology.orchestra.run.vm07.stdout:{ 2026-03-07T07:52:49.144 INFO:teuthology.orchestra.run.vm07.stdout: "ceph_version": "ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable)", 2026-03-07T07:52:49.144 INFO:teuthology.orchestra.run.vm07.stdout: "image_id": "8bccc98d839aa18345ec1336292d0452ca331737e49f12524f635044dcabcfe1", 2026-03-07T07:52:49.144 INFO:teuthology.orchestra.run.vm07.stdout: "repo_digests": [ 2026-03-07T07:52:49.144 INFO:teuthology.orchestra.run.vm07.stdout: "harbor.clyso.com/custom-ceph/ceph/ceph@sha256:ffa52c72fad7bdd2657408de9cf8d87fc2c72f716d1a00277ba13f7c12b404e0" 2026-03-07T07:52:49.144 INFO:teuthology.orchestra.run.vm07.stdout: ] 2026-03-07T07:52:49.144 INFO:teuthology.orchestra.run.vm07.stdout:} 2026-03-07T07:52:49.162 DEBUG:teuthology.orchestra.run.vm07:> sudo mkdir -p /etc/ceph 2026-03-07T07:52:49.170 DEBUG:teuthology.orchestra.run.vm09:> sudo mkdir -p /etc/ceph 2026-03-07T07:52:49.177 DEBUG:teuthology.orchestra.run.vm07:> sudo chmod 777 /etc/ceph 2026-03-07T07:52:49.219 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod 777 /etc/ceph 2026-03-07T07:52:49.227 INFO:tasks.cephadm:Writing seed config... 2026-03-07T07:52:49.228 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-07T07:52:49.228 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-07T07:52:49.228 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-07T07:52:49.228 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-07T07:52:49.228 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-07T07:52:49.228 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-07T07:52:49.228 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-07T07:52:49.228 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-07T07:52:49.228 INFO:tasks.cephadm: override: [osd] osd shutdown pgref assert = True 2026-03-07T07:52:49.228 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-07T07:52:49.228 DEBUG:teuthology.orchestra.run.vm07:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-07T07:52:49.267 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = True bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-07T07:52:49.267 DEBUG:teuthology.orchestra.run.vm07:mon.vm07> sudo journalctl -f -n 0 -u ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@mon.vm07.service 2026-03-07T07:52:49.308 INFO:tasks.cephadm:Bootstrapping... 2026-03-07T07:52:49.308 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 -v bootstrap --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-ip 192.168.123.107 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-07T07:52:49.590 INFO:teuthology.orchestra.run.vm07.stdout:-------------------------------------------------------------------------------- 2026-03-07T07:52:49.590 INFO:teuthology.orchestra.run.vm07.stdout:cephadm ['--image', 'harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5', '-v', 'bootstrap', '--fsid', '312fdbc4-19f2-11f1-81d0-4bbd10a1e012', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-ip', '192.168.123.107', '--skip-admin-label'] 2026-03-07T07:52:49.590 INFO:teuthology.orchestra.run.vm07.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-07T07:52:49.591 INFO:teuthology.orchestra.run.vm07.stdout:Verifying podman|docker is present... 2026-03-07T07:52:49.591 INFO:teuthology.orchestra.run.vm07.stdout:Verifying lvm2 is present... 2026-03-07T07:52:49.591 INFO:teuthology.orchestra.run.vm07.stdout:Verifying time synchronization is in place... 2026-03-07T07:52:49.593 INFO:teuthology.orchestra.run.vm07.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-07T07:52:49.593 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-07T07:52:49.595 INFO:teuthology.orchestra.run.vm07.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-07T07:52:49.595 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stdout inactive 2026-03-07T07:52:49.597 INFO:teuthology.orchestra.run.vm07.stdout:Non-zero exit code 1 from systemctl is-enabled chronyd.service 2026-03-07T07:52:49.598 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stderr Failed to get unit file state for chronyd.service: No such file or directory 2026-03-07T07:52:49.600 INFO:teuthology.orchestra.run.vm07.stdout:Non-zero exit code 3 from systemctl is-active chronyd.service 2026-03-07T07:52:49.600 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stdout inactive 2026-03-07T07:52:49.602 INFO:teuthology.orchestra.run.vm07.stdout:Non-zero exit code 1 from systemctl is-enabled systemd-timesyncd.service 2026-03-07T07:52:49.602 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stdout masked 2026-03-07T07:52:49.604 INFO:teuthology.orchestra.run.vm07.stdout:Non-zero exit code 3 from systemctl is-active systemd-timesyncd.service 2026-03-07T07:52:49.604 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stdout inactive 2026-03-07T07:52:49.607 INFO:teuthology.orchestra.run.vm07.stdout:Non-zero exit code 1 from systemctl is-enabled ntpd.service 2026-03-07T07:52:49.607 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stderr Failed to get unit file state for ntpd.service: No such file or directory 2026-03-07T07:52:49.609 INFO:teuthology.orchestra.run.vm07.stdout:Non-zero exit code 3 from systemctl is-active ntpd.service 2026-03-07T07:52:49.609 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stdout inactive 2026-03-07T07:52:49.612 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stdout enabled 2026-03-07T07:52:49.614 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stdout active 2026-03-07T07:52:49.614 INFO:teuthology.orchestra.run.vm07.stdout:Unit ntp.service is enabled and running 2026-03-07T07:52:49.614 INFO:teuthology.orchestra.run.vm07.stdout:Repeating the final host check... 2026-03-07T07:52:49.614 INFO:teuthology.orchestra.run.vm07.stdout:docker (/usr/bin/docker) is present 2026-03-07T07:52:49.614 INFO:teuthology.orchestra.run.vm07.stdout:systemctl is present 2026-03-07T07:52:49.614 INFO:teuthology.orchestra.run.vm07.stdout:lvcreate is present 2026-03-07T07:52:49.616 INFO:teuthology.orchestra.run.vm07.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-07T07:52:49.616 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-07T07:52:49.618 INFO:teuthology.orchestra.run.vm07.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-07T07:52:49.618 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stdout inactive 2026-03-07T07:52:49.620 INFO:teuthology.orchestra.run.vm07.stdout:Non-zero exit code 1 from systemctl is-enabled chronyd.service 2026-03-07T07:52:49.620 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stderr Failed to get unit file state for chronyd.service: No such file or directory 2026-03-07T07:52:49.622 INFO:teuthology.orchestra.run.vm07.stdout:Non-zero exit code 3 from systemctl is-active chronyd.service 2026-03-07T07:52:49.622 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stdout inactive 2026-03-07T07:52:49.624 INFO:teuthology.orchestra.run.vm07.stdout:Non-zero exit code 1 from systemctl is-enabled systemd-timesyncd.service 2026-03-07T07:52:49.624 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stdout masked 2026-03-07T07:52:49.626 INFO:teuthology.orchestra.run.vm07.stdout:Non-zero exit code 3 from systemctl is-active systemd-timesyncd.service 2026-03-07T07:52:49.626 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stdout inactive 2026-03-07T07:52:49.628 INFO:teuthology.orchestra.run.vm07.stdout:Non-zero exit code 1 from systemctl is-enabled ntpd.service 2026-03-07T07:52:49.628 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stderr Failed to get unit file state for ntpd.service: No such file or directory 2026-03-07T07:52:49.630 INFO:teuthology.orchestra.run.vm07.stdout:Non-zero exit code 3 from systemctl is-active ntpd.service 2026-03-07T07:52:49.630 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stdout inactive 2026-03-07T07:52:49.633 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stdout enabled 2026-03-07T07:52:49.635 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stdout active 2026-03-07T07:52:49.635 INFO:teuthology.orchestra.run.vm07.stdout:Unit ntp.service is enabled and running 2026-03-07T07:52:49.635 INFO:teuthology.orchestra.run.vm07.stdout:Host looks OK 2026-03-07T07:52:49.635 INFO:teuthology.orchestra.run.vm07.stdout:Cluster fsid: 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:52:49.635 INFO:teuthology.orchestra.run.vm07.stdout:Acquiring lock 139891683330224 on /run/cephadm/312fdbc4-19f2-11f1-81d0-4bbd10a1e012.lock 2026-03-07T07:52:49.635 INFO:teuthology.orchestra.run.vm07.stdout:Lock 139891683330224 acquired on /run/cephadm/312fdbc4-19f2-11f1-81d0-4bbd10a1e012.lock 2026-03-07T07:52:49.636 INFO:teuthology.orchestra.run.vm07.stdout:Verifying IP 192.168.123.107 port 3300 ... 2026-03-07T07:52:49.636 INFO:teuthology.orchestra.run.vm07.stdout:Verifying IP 192.168.123.107 port 6789 ... 2026-03-07T07:52:49.636 INFO:teuthology.orchestra.run.vm07.stdout:Base mon IP(s) is [192.168.123.107:3300, 192.168.123.107:6789], mon addrv is [v2:192.168.123.107:3300,v1:192.168.123.107:6789] 2026-03-07T07:52:49.637 INFO:teuthology.orchestra.run.vm07.stdout:/usr/sbin/ip: stdout default via 192.168.123.1 dev ens3 proto dhcp src 192.168.123.107 metric 100 2026-03-07T07:52:49.637 INFO:teuthology.orchestra.run.vm07.stdout:/usr/sbin/ip: stdout 172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.0.1 linkdown 2026-03-07T07:52:49.637 INFO:teuthology.orchestra.run.vm07.stdout:/usr/sbin/ip: stdout 192.168.123.0/24 dev ens3 proto kernel scope link src 192.168.123.107 metric 100 2026-03-07T07:52:49.637 INFO:teuthology.orchestra.run.vm07.stdout:/usr/sbin/ip: stdout 192.168.123.1 dev ens3 proto dhcp scope link src 192.168.123.107 metric 100 2026-03-07T07:52:49.639 INFO:teuthology.orchestra.run.vm07.stdout:/usr/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-07T07:52:49.639 INFO:teuthology.orchestra.run.vm07.stdout:/usr/sbin/ip: stdout fe80::/64 dev ens3 proto kernel metric 256 pref medium 2026-03-07T07:52:49.640 INFO:teuthology.orchestra.run.vm07.stdout:/usr/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-07T07:52:49.640 INFO:teuthology.orchestra.run.vm07.stdout:/usr/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-07T07:52:49.640 INFO:teuthology.orchestra.run.vm07.stdout:/usr/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-07T07:52:49.640 INFO:teuthology.orchestra.run.vm07.stdout:/usr/sbin/ip: stdout 2: ens3: mtu 1500 state UP qlen 1000 2026-03-07T07:52:49.640 INFO:teuthology.orchestra.run.vm07.stdout:/usr/sbin/ip: stdout inet6 fe80::5055:ff:fe00:7/64 scope link 2026-03-07T07:52:49.640 INFO:teuthology.orchestra.run.vm07.stdout:/usr/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-07T07:52:49.641 INFO:teuthology.orchestra.run.vm07.stdout:Mon IP `192.168.123.107` is in CIDR network `192.168.123.0/24` 2026-03-07T07:52:49.641 INFO:teuthology.orchestra.run.vm07.stdout:Mon IP `192.168.123.107` is in CIDR network `192.168.123.0/24` 2026-03-07T07:52:49.641 INFO:teuthology.orchestra.run.vm07.stdout:Mon IP `192.168.123.107` is in CIDR network `192.168.123.1/32` 2026-03-07T07:52:49.641 INFO:teuthology.orchestra.run.vm07.stdout:Mon IP `192.168.123.107` is in CIDR network `192.168.123.1/32` 2026-03-07T07:52:49.641 INFO:teuthology.orchestra.run.vm07.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24', '192.168.123.1/32', '192.168.123.1/32'] 2026-03-07T07:52:49.641 INFO:teuthology.orchestra.run.vm07.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-07T07:52:49.642 INFO:teuthology.orchestra.run.vm07.stdout:Pulling container image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5... 2026-03-07T07:52:50.168 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/docker: stdout cobaltcore-storage-v19.2.3-fasttrack-5: Pulling from custom-ceph/ceph/ceph 2026-03-07T07:52:50.169 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/docker: stdout Digest: sha256:ffa52c72fad7bdd2657408de9cf8d87fc2c72f716d1a00277ba13f7c12b404e0 2026-03-07T07:52:50.169 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/docker: stdout Status: Image is up to date for harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 2026-03-07T07:52:50.169 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/docker: stdout harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 2026-03-07T07:52:50.419 INFO:teuthology.orchestra.run.vm07.stdout:ceph: stdout ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable) 2026-03-07T07:52:50.419 INFO:teuthology.orchestra.run.vm07.stdout:Ceph version: ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable) 2026-03-07T07:52:50.419 INFO:teuthology.orchestra.run.vm07.stdout:Extracting ceph user uid/gid from container image... 2026-03-07T07:52:50.519 INFO:teuthology.orchestra.run.vm07.stdout:stat: stdout 167 167 2026-03-07T07:52:50.519 INFO:teuthology.orchestra.run.vm07.stdout:Creating initial keys... 2026-03-07T07:52:50.616 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-authtool: stdout AQBCy6tp7mPlIhAAwfCpxsutM6b3FHTxTIx0nA== 2026-03-07T07:52:50.711 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-authtool: stdout AQBCy6tpZz+hKBAAjFbbpPnJmSIeFcRwzNxS+w== 2026-03-07T07:52:50.808 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-authtool: stdout AQBCy6tpcjMKLhAAPyVZLlNEc3WkUxHRJW206A== 2026-03-07T07:52:50.808 INFO:teuthology.orchestra.run.vm07.stdout:Creating initial monmap... 2026-03-07T07:52:50.903 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-07T07:52:50.903 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-07T07:52:50.903 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:52:50.903 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-07T07:52:50.903 INFO:teuthology.orchestra.run.vm07.stdout:monmaptool for vm07 [v2:192.168.123.107:3300,v1:192.168.123.107:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-07T07:52:50.903 INFO:teuthology.orchestra.run.vm07.stdout:setting min_mon_release = quincy 2026-03-07T07:52:50.903 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/monmaptool: set fsid to 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:52:50.903 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-07T07:52:50.903 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:52:50.903 INFO:teuthology.orchestra.run.vm07.stdout:Creating mon... 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.963+0000 7fcb006bbd80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.963+0000 7fcb006bbd80 1 imported monmap: 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr epoch 0 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr last_changed 2026-03-07T06:52:50.872627+0000 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr created 2026-03-07T06:52:50.872627+0000 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr min_mon_release 17 (quincy) 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr election_strategy: 1 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 0: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.vm07 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.963+0000 7fcb006bbd80 0 /usr/bin/ceph-mon: set fsid to 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: RocksDB version: 7.9.2 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Git sha 0 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Compile date 2026-03-06 13:52:12 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: DB SUMMARY 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: DB Session ID: S4V0O44SC3MLDSQZT4QL 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-vm07/store.db dir, Total Num: 0, files: 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-vm07/store.db: 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.error_if_exists: 0 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.create_if_missing: 1 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.paranoid_checks: 1 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.env: 0x5613c9361ca0 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.info_log: 0x5613d5d28ce0 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.statistics: (nil) 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.use_fsync: 0 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_log_file_size: 0 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.allow_fallocate: 1 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.use_direct_reads: 0 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-07T07:52:51.020 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.db_log_dir: 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.wal_dir: 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.write_buffer_manager: 0x5613d5d1f5e0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.unordered_write: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.row_cache: None 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.wal_filter: None 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.two_write_queues: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.wal_compression: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.atomic_flush: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.log_readahead_size: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_background_jobs: 2 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_background_compactions: -1 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_subcompactions: 1 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_open_files: -1 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_background_flushes: -1 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Compression algorithms supported: 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: kZSTD supported: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: kXpressCompression supported: 0 2026-03-07T07:52:51.021 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: kBZip2Compression supported: 0 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: kLZ4Compression supported: 1 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: kZlibCompression supported: 1 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: kSnappyCompression supported: 1 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: [db/db_impl/db_impl_open.cc:317] Creating manifest 1 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-vm07/store.db/MANIFEST-000001 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.merge_operator: 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compaction_filter: None 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5613d5d1b400) 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr cache_index_and_filter_blocks: 1 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr cache_index_and_filter_blocks_with_high_priority: 0 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr pin_top_level_index_and_filter: 1 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr index_type: 0 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr data_block_index_type: 0 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr index_shortening: 1 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr data_block_hash_table_util_ratio: 0.750000 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr checksum: 4 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr no_block_cache: 0 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr block_cache: 0x5613d5d411f0 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr block_cache_name: BinnedLRUCache 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr block_cache_options: 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr capacity : 536870912 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr num_shard_bits : 4 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr strict_capacity_limit : 0 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr high_pri_pool_ratio: 0.000 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr block_cache_compressed: (nil) 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr persistent_cache: (nil) 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr block_size: 4096 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr block_size_deviation: 10 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr block_restart_interval: 16 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr index_block_restart_interval: 1 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr metadata_block_size: 4096 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr partition_filters: 0 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr use_delta_encoding: 1 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr filter_policy: bloomfilter 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr whole_key_filtering: 1 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr verify_compression: 0 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr read_amp_bytes_per_bit: 0 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr format_version: 5 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr enable_index_compression: 1 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr block_align: 0 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr max_auto_readahead_size: 262144 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr prepopulate_block_cache: 0 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr initial_auto_readahead_size: 8192 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr num_file_reads_for_auto_readahead: 2 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compression: NoCompression 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.num_levels: 7 2026-03-07T07:52:51.022 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.inplace_update_support: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.bloom_locality: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.max_successive_merges: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.ttl: 2592000 2026-03-07T07:52:51.023 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.enable_blob_files: false 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.min_blob_size: 0 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.967+0000 7fcb006bbd80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.971+0000 7fcb006bbd80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-vm07/store.db/MANIFEST-000001 succeeded,manifest_file_number is 1, next_file_number is 3, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.971+0000 7fcb006bbd80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.971+0000 7fcb006bbd80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 5c30debd-8448-4f39-babe-01334d278f74 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.971+0000 7fcb006bbd80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 5 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.971+0000 7fcb006bbd80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x5613d5d42e00 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.971+0000 7fcb006bbd80 4 rocksdb: DB pointer 0x5613d5e26000 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.975+0000 7fcaf7e45640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.975+0000 7fcaf7e45640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr ** DB Stats ** 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr Uptime(secs): 0.0 total, 0.0 interval 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr ** Compaction Stats [default] ** 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr ** Compaction Stats [default] ** 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr Uptime(secs): 0.0 total, 0.0 interval 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr Flush(GB): cumulative 0.000, interval 0.000 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr AddFile(GB): cumulative 0.000, interval 0.000 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr AddFile(Total Files): cumulative 0, interval 0 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr AddFile(L0 Files): cumulative 0, interval 0 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr AddFile(Keys): cumulative 0, interval 0 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr Block cache BinnedLRUCache@0x5613d5d411f0#7 capacity: 512.00 MB usage: 0.00 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 7e-06 secs_since: 0 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr Block cache entry stats(count,size,portion): Misc(1,0.00 KB,0%) 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr ** File Read Latency Histogram By Level [default] ** 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.975+0000 7fcb006bbd80 4 rocksdb: [db/db_impl/db_impl.cc:496] Shutdown: canceling all background work 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.975+0000 7fcb006bbd80 4 rocksdb: [db/db_impl/db_impl.cc:704] Shutdown complete 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-07T06:52:50.975+0000 7fcb006bbd80 0 /usr/bin/ceph-mon: created monfs at /var/lib/ceph/mon/ceph-vm07 for mon.vm07 2026-03-07T07:52:51.024 INFO:teuthology.orchestra.run.vm07.stdout:create mon.vm07 on 2026-03-07T07:52:51.362 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-07T07:52:51.559 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012.target → /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012.target. 2026-03-07T07:52:51.559 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012.target → /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012.target. 2026-03-07T07:52:51.753 INFO:teuthology.orchestra.run.vm07.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@mon.vm07 2026-03-07T07:52:51.753 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stderr Failed to reset failed state of unit ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@mon.vm07.service: Unit ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@mon.vm07.service not loaded. 2026-03-07T07:52:51.904 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012.target.wants/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@mon.vm07.service → /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service. 2026-03-07T07:52:51.909 INFO:teuthology.orchestra.run.vm07.stdout:firewalld does not appear to be present 2026-03-07T07:52:51.910 INFO:teuthology.orchestra.run.vm07.stdout:Not possible to enable service . firewalld.service is not available 2026-03-07T07:52:51.910 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for mon to start... 2026-03-07T07:52:51.910 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for mon... 2026-03-07T07:52:52.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:52 vm07 bash[16565]: cluster 2026-03-07T06:52:52.034274+0000 mon.vm07 (mon.0) 1 : cluster [INF] mon.vm07 is new leader, mons vm07 in quorum (ranks 0) 2026-03-07T07:52:52.227 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout cluster: 2026-03-07T07:52:52.227 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout id: 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:52:52.227 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-07T07:52:52.227 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 2026-03-07T07:52:52.227 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout services: 2026-03-07T07:52:52.227 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum vm07 (age 0.137083s) 2026-03-07T07:52:52.227 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-07T07:52:52.227 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-07T07:52:52.227 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 2026-03-07T07:52:52.227 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout data: 2026-03-07T07:52:52.227 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-07T07:52:52.227 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-07T07:52:52.227 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-07T07:52:52.227 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout pgs: 2026-03-07T07:52:52.227 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 2026-03-07T07:52:52.227 INFO:teuthology.orchestra.run.vm07.stdout:mon is available 2026-03-07T07:52:52.227 INFO:teuthology.orchestra.run.vm07.stdout:Assimilating anything we can from ceph.conf... 2026-03-07T07:52:52.546 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 2026-03-07T07:52:52.546 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout [global] 2026-03-07T07:52:52.546 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout fsid = 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:52:52.546 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-07T07:52:52.546 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.107:3300,v1:192.168.123.107:6789] 2026-03-07T07:52:52.546 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-07T07:52:52.546 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-07T07:52:52.546 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-07T07:52:52.546 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-07T07:52:52.546 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 2026-03-07T07:52:52.546 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-07T07:52:52.546 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-07T07:52:52.546 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 2026-03-07T07:52:52.546 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout [osd] 2026-03-07T07:52:52.546 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-07T07:52:52.546 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-07T07:52:52.546 INFO:teuthology.orchestra.run.vm07.stdout:Generating new minimal ceph.conf... 2026-03-07T07:52:52.843 INFO:teuthology.orchestra.run.vm07.stdout:Restarting the monitor... 2026-03-07T07:52:53.004 INFO:teuthology.orchestra.run.vm07.stdout:Setting public_network to 192.168.123.0/24,192.168.123.1/32 in mon config section 2026-03-07T07:52:53.054 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:52 vm07 systemd[1]: Stopping Ceph mon.vm07 for 312fdbc4-19f2-11f1-81d0-4bbd10a1e012... 2026-03-07T07:52:53.054 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:52 vm07 bash[16565]: debug 2026-03-07T06:52:52.871+0000 7fa72e613640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.vm07 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-07T07:52:53.054 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:52 vm07 bash[16565]: debug 2026-03-07T06:52:52.871+0000 7fa72e613640 -1 mon.vm07@0(leader) e1 *** Got Signal Terminated *** 2026-03-07T07:52:53.054 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:52 vm07 bash[16943]: ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012-mon-vm07 2026-03-07T07:52:53.054 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:52 vm07 systemd[1]: ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@mon.vm07.service: Deactivated successfully. 2026-03-07T07:52:53.054 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:52 vm07 systemd[1]: Stopped Ceph mon.vm07 for 312fdbc4-19f2-11f1-81d0-4bbd10a1e012. 2026-03-07T07:52:53.054 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:52 vm07 systemd[1]: Started Ceph mon.vm07 for 312fdbc4-19f2-11f1-81d0-4bbd10a1e012. 2026-03-07T07:52:53.313 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.123+0000 7f388d257d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.123+0000 7f388d257d80 0 ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable), process ceph-mon, pid 7 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.123+0000 7f388d257d80 0 pidfile_write: ignore empty --pid-file 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 0 load: jerasure load: lrc 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Git sha 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Compile date 2026-03-06 13:52:12 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: DB SUMMARY 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: DB Session ID: YUX0XF7Z3E0E7EOPNZB2 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: CURRENT file: CURRENT 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-vm07/store.db dir, Total Num: 1, files: 000008.sst 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-vm07/store.db: 000009.log size: 75215 ; 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.error_if_exists: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.create_if_missing: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.env: 0x55bc9a0e6ca0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.info_log: 0x55bca82d0500 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.statistics: (nil) 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.use_fsync: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.db_log_dir: 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.wal_dir: 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.write_buffer_manager: 0x55bca82d5900 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.unordered_write: 0 2026-03-07T07:52:53.314 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.row_cache: None 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.wal_filter: None 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.two_write_queues: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.wal_compression: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.atomic_flush: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_open_files: -1 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Compression algorithms supported: 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: kZSTD supported: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: kXpressCompression supported: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: kZlibCompression supported: 1 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-vm07/store.db/MANIFEST-000010 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.merge_operator: 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compaction_filter: None 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55bca82d04c0) 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cache_index_and_filter_blocks: 1 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-07T07:52:53.315 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: pin_top_level_index_and_filter: 1 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: index_type: 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: data_block_index_type: 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: index_shortening: 1 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: data_block_hash_table_util_ratio: 0.750000 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: checksum: 4 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: no_block_cache: 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: block_cache: 0x55bca82f71f0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: block_cache_name: BinnedLRUCache 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: block_cache_options: 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: capacity : 536870912 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: num_shard_bits : 4 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: strict_capacity_limit : 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: high_pri_pool_ratio: 0.000 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: block_cache_compressed: (nil) 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: persistent_cache: (nil) 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: block_size: 4096 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: block_size_deviation: 10 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: block_restart_interval: 16 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: index_block_restart_interval: 1 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: metadata_block_size: 4096 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: partition_filters: 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: use_delta_encoding: 1 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: filter_policy: bloomfilter 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: whole_key_filtering: 1 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: verify_compression: 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: read_amp_bytes_per_bit: 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: format_version: 5 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: enable_index_compression: 1 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: block_align: 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: max_auto_readahead_size: 262144 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: prepopulate_block_cache: 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: initial_auto_readahead_size: 8192 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: num_file_reads_for_auto_readahead: 2 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compression: NoCompression 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.num_levels: 7 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-07T07:52:53.316 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.bloom_locality: 0 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.ttl: 2592000 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.enable_blob_files: false 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.min_blob_size: 0 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-07T07:52:53.317 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.127+0000 7f388d257d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.139+0000 7f388d257d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-vm07/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.139+0000 7f388d257d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.139+0000 7f388d257d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 5c30debd-8448-4f39-babe-01334d278f74 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.139+0000 7f388d257d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772866373143732, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.139+0000 7f388d257d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.139+0000 7f388d257d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772866373145636, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 72311, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 8, "largest_seqno": 223, "table_properties": {"data_size": 70593, "index_size": 171, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 517, "raw_key_size": 9562, "raw_average_key_size": 49, "raw_value_size": 65187, "raw_average_value_size": 336, "num_data_blocks": 8, "num_entries": 194, "num_filter_entries": 194, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1772866373, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "5c30debd-8448-4f39-babe-01334d278f74", "db_session_id": "YUX0XF7Z3E0E7EOPNZB2", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.139+0000 7f388d257d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772866373145699, "job": 1, "event": "recovery_finished"} 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.139+0000 7f388d257d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.143+0000 7f388d257d80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-vm07/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.143+0000 7f388d257d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55bca82f8e00 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.143+0000 7f388d257d80 4 rocksdb: DB pointer 0x55bca8414000 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.143+0000 7f3883021640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: debug 2026-03-07T06:52:53.143+0000 7f3883021640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: ** DB Stats ** 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: ** Compaction Stats [default] ** 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: L0 2/0 72.49 KB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 40.7 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: Sum 2/0 72.49 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 40.7 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 40.7 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: ** Compaction Stats [default] ** 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 40.7 0.00 0.00 1 0.002 0 0 0.0 0.0 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: Flush(GB): cumulative 0.000, interval 0.000 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: AddFile(Total Files): cumulative 0, interval 0 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: AddFile(Keys): cumulative 0, interval 0 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: Cumulative compaction: 0.00 GB write, 4.04 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: Interval compaction: 0.00 GB write, 4.04 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: Block cache BinnedLRUCache@0x55bca82f71f0#7 capacity: 512.00 MB usage: 1.06 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1e-05 secs_since: 0 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: Block cache entry stats(count,size,portion): FilterBlock(2,0.70 KB,0.00013411%) IndexBlock(2,0.36 KB,6.85453e-05%) Misc(1,0.00 KB,0%) 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: ** File Read Latency Histogram By Level [default] ** 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.155823+0000 mon.vm07 (mon.0) 1 : cluster [INF] mon.vm07 is new leader, mons vm07 in quorum (ranks 0) 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.155823+0000 mon.vm07 (mon.0) 1 : cluster [INF] mon.vm07 is new leader, mons vm07 in quorum (ranks 0) 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.155857+0000 mon.vm07 (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.155857+0000 mon.vm07 (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.155862+0000 mon.vm07 (mon.0) 3 : cluster [DBG] fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.155862+0000 mon.vm07 (mon.0) 3 : cluster [DBG] fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.155866+0000 mon.vm07 (mon.0) 4 : cluster [DBG] last_changed 2026-03-07T06:52:50.872627+0000 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.155866+0000 mon.vm07 (mon.0) 4 : cluster [DBG] last_changed 2026-03-07T06:52:50.872627+0000 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.155874+0000 mon.vm07 (mon.0) 5 : cluster [DBG] created 2026-03-07T06:52:50.872627+0000 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.155874+0000 mon.vm07 (mon.0) 5 : cluster [DBG] created 2026-03-07T06:52:50.872627+0000 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.155879+0000 mon.vm07 (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.155879+0000 mon.vm07 (mon.0) 6 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.155885+0000 mon.vm07 (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.155885+0000 mon.vm07 (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.155889+0000 mon.vm07 (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.vm07 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.155889+0000 mon.vm07 (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.vm07 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.156107+0000 mon.vm07 (mon.0) 9 : cluster [DBG] fsmap 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.156107+0000 mon.vm07 (mon.0) 9 : cluster [DBG] fsmap 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.156124+0000 mon.vm07 (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.156124+0000 mon.vm07 (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-07T07:52:53.318 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.156622+0000 mon.vm07 (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-07T07:52:53.319 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 bash[17031]: cluster 2026-03-07T06:52:53.156622+0000 mon.vm07 (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-07T07:52:53.347 INFO:teuthology.orchestra.run.vm07.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-07T07:52:53.348 INFO:teuthology.orchestra.run.vm07.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-07T07:52:53.348 INFO:teuthology.orchestra.run.vm07.stdout:Creating mgr... 2026-03-07T07:52:53.348 INFO:teuthology.orchestra.run.vm07.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-07T07:52:53.348 INFO:teuthology.orchestra.run.vm07.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-07T07:52:53.348 INFO:teuthology.orchestra.run.vm07.stdout:Verifying port 0.0.0.0:8443 ... 2026-03-07T07:52:53.516 INFO:teuthology.orchestra.run.vm07.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@mgr.vm07.yrfcuj 2026-03-07T07:52:53.517 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stderr Failed to reset failed state of unit ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@mgr.vm07.yrfcuj.service: Unit ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@mgr.vm07.yrfcuj.service not loaded. 2026-03-07T07:52:53.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:52:53.707 INFO:teuthology.orchestra.run.vm07.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012.target.wants/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@mgr.vm07.yrfcuj.service → /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service. 2026-03-07T07:52:53.715 INFO:teuthology.orchestra.run.vm07.stdout:firewalld does not appear to be present 2026-03-07T07:52:53.715 INFO:teuthology.orchestra.run.vm07.stdout:Not possible to enable service . firewalld.service is not available 2026-03-07T07:52:53.715 INFO:teuthology.orchestra.run.vm07.stdout:firewalld does not appear to be present 2026-03-07T07:52:53.715 INFO:teuthology.orchestra.run.vm07.stdout:Not possible to open ports <[9283, 8765, 8443]>. firewalld.service is not available 2026-03-07T07:52:53.715 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for mgr to start... 2026-03-07T07:52:53.715 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for mgr... 2026-03-07T07:52:53.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:53 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout { 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "fsid": "312fdbc4-19f2-11f1-81d0-4bbd10a1e012", 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "health": { 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 0 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout ], 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "vm07" 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout ], 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "quorum_age": 0, 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-07T07:52:54.059 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "btime": "2026-03-07T06:52:52:038351+0000", 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "restful" 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout ], 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "modified": "2026-03-07T06:52:52.038926+0000", 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout } 2026-03-07T07:52:54.060 INFO:teuthology.orchestra.run.vm07.stdout:mgr not available, waiting (1/15)... 2026-03-07T07:52:54.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:54 vm07 bash[17031]: audit 2026-03-07T06:52:53.297980+0000 mon.vm07 (mon.0) 12 : audit [INF] from='client.? 192.168.123.107:0/3958703610' entity='client.admin' 2026-03-07T07:52:54.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:54 vm07 bash[17031]: audit 2026-03-07T06:52:53.297980+0000 mon.vm07 (mon.0) 12 : audit [INF] from='client.? 192.168.123.107:0/3958703610' entity='client.admin' 2026-03-07T07:52:54.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:54 vm07 bash[17031]: audit 2026-03-07T06:52:53.995885+0000 mon.vm07 (mon.0) 13 : audit [DBG] from='client.? 192.168.123.107:0/702670522' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T07:52:54.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:54 vm07 bash[17031]: audit 2026-03-07T06:52:53.995885+0000 mon.vm07 (mon.0) 13 : audit [DBG] from='client.? 192.168.123.107:0/702670522' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T07:52:56.476 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 2026-03-07T07:52:56.476 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout { 2026-03-07T07:52:56.476 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "fsid": "312fdbc4-19f2-11f1-81d0-4bbd10a1e012", 2026-03-07T07:52:56.476 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "health": { 2026-03-07T07:52:56.476 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-07T07:52:56.476 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-07T07:52:56.476 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-07T07:52:56.476 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:56.476 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 0 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout ], 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "vm07" 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout ], 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "quorum_age": 3, 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-07T07:52:56.477 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "btime": "2026-03-07T06:52:52:038351+0000", 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "restful" 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout ], 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "modified": "2026-03-07T06:52:52.038926+0000", 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout } 2026-03-07T07:52:56.478 INFO:teuthology.orchestra.run.vm07.stdout:mgr not available, waiting (2/15)... 2026-03-07T07:52:56.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:56 vm07 bash[17031]: audit 2026-03-07T06:52:56.369738+0000 mon.vm07 (mon.0) 14 : audit [DBG] from='client.? 192.168.123.107:0/1211063319' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T07:52:56.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:56 vm07 bash[17031]: audit 2026-03-07T06:52:56.369738+0000 mon.vm07 (mon.0) 14 : audit [DBG] from='client.? 192.168.123.107:0/1211063319' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout { 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "fsid": "312fdbc4-19f2-11f1-81d0-4bbd10a1e012", 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "health": { 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 0 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout ], 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "vm07" 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout ], 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "quorum_age": 5, 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-07T07:52:58.797 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "btime": "2026-03-07T06:52:52:038351+0000", 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "restful" 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout ], 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "modified": "2026-03-07T06:52:52.038926+0000", 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout } 2026-03-07T07:52:58.798 INFO:teuthology.orchestra.run.vm07.stdout:mgr not available, waiting (3/15)... 2026-03-07T07:52:59.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:58 vm07 bash[17031]: audit 2026-03-07T06:52:58.746479+0000 mon.vm07 (mon.0) 15 : audit [DBG] from='client.? 192.168.123.107:0/129046028' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T07:52:59.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:52:58 vm07 bash[17031]: audit 2026-03-07T06:52:58.746479+0000 mon.vm07 (mon.0) 15 : audit [DBG] from='client.? 192.168.123.107:0/129046028' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout { 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "fsid": "312fdbc4-19f2-11f1-81d0-4bbd10a1e012", 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "health": { 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 0 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout ], 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "vm07" 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout ], 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "quorum_age": 7, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "btime": "2026-03-07T06:52:52:038351+0000", 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-07T07:53:01.180 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-07T07:53:01.181 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-07T07:53:01.181 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-07T07:53:01.181 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-07T07:53:01.181 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "restful" 2026-03-07T07:53:01.181 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout ], 2026-03-07T07:53:01.181 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T07:53:01.181 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:01.181 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-07T07:53:01.181 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:53:01.181 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "modified": "2026-03-07T06:52:52.038926+0000", 2026-03-07T07:53:01.181 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T07:53:01.181 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:01.181 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-07T07:53:01.181 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout } 2026-03-07T07:53:01.181 INFO:teuthology.orchestra.run.vm07.stdout:mgr not available, waiting (4/15)... 2026-03-07T07:53:01.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:01 vm07 bash[17031]: audit 2026-03-07T06:53:01.114614+0000 mon.vm07 (mon.0) 16 : audit [DBG] from='client.? 192.168.123.107:0/1678834000' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T07:53:01.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:01 vm07 bash[17031]: audit 2026-03-07T06:53:01.114614+0000 mon.vm07 (mon.0) 16 : audit [DBG] from='client.? 192.168.123.107:0/1678834000' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T07:53:03.650 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 2026-03-07T07:53:03.650 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout { 2026-03-07T07:53:03.650 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "fsid": "312fdbc4-19f2-11f1-81d0-4bbd10a1e012", 2026-03-07T07:53:03.650 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "health": { 2026-03-07T07:53:03.650 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-07T07:53:03.650 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-07T07:53:03.650 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 0 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout ], 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "vm07" 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout ], 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "quorum_age": 10, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "btime": "2026-03-07T06:52:52:038351+0000", 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "restful" 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout ], 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "modified": "2026-03-07T06:52:52.038926+0000", 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout } 2026-03-07T07:53:03.651 INFO:teuthology.orchestra.run.vm07.stdout:mgr not available, waiting (5/15)... 2026-03-07T07:53:03.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:03 vm07 bash[17031]: audit 2026-03-07T06:53:03.575446+0000 mon.vm07 (mon.0) 17 : audit [DBG] from='client.? 192.168.123.107:0/95912206' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T07:53:03.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:03 vm07 bash[17031]: audit 2026-03-07T06:53:03.575446+0000 mon.vm07 (mon.0) 17 : audit [DBG] from='client.? 192.168.123.107:0/95912206' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T07:53:04.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: cluster 2026-03-07T06:53:03.648222+0000 mon.vm07 (mon.0) 18 : cluster [INF] Activating manager daemon vm07.yrfcuj 2026-03-07T07:53:04.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: cluster 2026-03-07T06:53:03.648222+0000 mon.vm07 (mon.0) 18 : cluster [INF] Activating manager daemon vm07.yrfcuj 2026-03-07T07:53:04.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: cluster 2026-03-07T06:53:03.652871+0000 mon.vm07 (mon.0) 19 : cluster [DBG] mgrmap e2: vm07.yrfcuj(active, starting, since 0.00481426s) 2026-03-07T07:53:04.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: cluster 2026-03-07T06:53:03.652871+0000 mon.vm07 (mon.0) 19 : cluster [DBG] mgrmap e2: vm07.yrfcuj(active, starting, since 0.00481426s) 2026-03-07T07:53:04.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.653821+0000 mon.vm07 (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-07T07:53:04.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.653821+0000 mon.vm07 (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-07T07:53:04.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.653883+0000 mon.vm07 (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-07T07:53:04.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.653883+0000 mon.vm07 (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-07T07:53:04.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.653927+0000 mon.vm07 (mon.0) 22 : audit [DBG] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-07T07:53:04.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.653927+0000 mon.vm07 (mon.0) 22 : audit [DBG] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-07T07:53:04.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.654197+0000 mon.vm07 (mon.0) 23 : audit [DBG] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:53:04.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.654197+0000 mon.vm07 (mon.0) 23 : audit [DBG] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:53:04.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.654273+0000 mon.vm07 (mon.0) 24 : audit [DBG] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr metadata", "who": "vm07.yrfcuj", "id": "vm07.yrfcuj"}]: dispatch 2026-03-07T07:53:04.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.654273+0000 mon.vm07 (mon.0) 24 : audit [DBG] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr metadata", "who": "vm07.yrfcuj", "id": "vm07.yrfcuj"}]: dispatch 2026-03-07T07:53:04.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: cluster 2026-03-07T06:53:03.658571+0000 mon.vm07 (mon.0) 25 : cluster [INF] Manager daemon vm07.yrfcuj is now available 2026-03-07T07:53:04.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: cluster 2026-03-07T06:53:03.658571+0000 mon.vm07 (mon.0) 25 : cluster [INF] Manager daemon vm07.yrfcuj is now available 2026-03-07T07:53:04.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.666609+0000 mon.vm07 (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/mirror_snapshot_schedule"}]: dispatch 2026-03-07T07:53:04.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.666609+0000 mon.vm07 (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/mirror_snapshot_schedule"}]: dispatch 2026-03-07T07:53:04.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.669840+0000 mon.vm07 (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:04.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.669840+0000 mon.vm07 (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:04.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.670057+0000 mon.vm07 (mon.0) 28 : audit [INF] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/trash_purge_schedule"}]: dispatch 2026-03-07T07:53:04.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.670057+0000 mon.vm07 (mon.0) 28 : audit [INF] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/trash_purge_schedule"}]: dispatch 2026-03-07T07:53:04.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.671904+0000 mon.vm07 (mon.0) 29 : audit [INF] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:04.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.671904+0000 mon.vm07 (mon.0) 29 : audit [INF] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:04.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.673587+0000 mon.vm07 (mon.0) 30 : audit [INF] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:04.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:04 vm07 bash[17031]: audit 2026-03-07T06:53:03.673587+0000 mon.vm07 (mon.0) 30 : audit [INF] from='mgr.14100 192.168.123.107:0/3151343023' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:06.032 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:05 vm07 bash[17031]: cluster 2026-03-07T06:53:04.656458+0000 mon.vm07 (mon.0) 31 : cluster [DBG] mgrmap e3: vm07.yrfcuj(active, since 1.0084s) 2026-03-07T07:53:06.032 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:05 vm07 bash[17031]: cluster 2026-03-07T06:53:04.656458+0000 mon.vm07 (mon.0) 31 : cluster [DBG] mgrmap e3: vm07.yrfcuj(active, since 1.0084s) 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout { 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "fsid": "312fdbc4-19f2-11f1-81d0-4bbd10a1e012", 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "health": { 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 0 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout ], 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "vm07" 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout ], 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "quorum_age": 12, 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-07T07:53:06.062 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "btime": "2026-03-07T06:52:52:038351+0000", 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "restful" 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout ], 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "modified": "2026-03-07T06:52:52.038926+0000", 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout }, 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout } 2026-03-07T07:53:06.063 INFO:teuthology.orchestra.run.vm07.stdout:mgr is available 2026-03-07T07:53:06.411 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 2026-03-07T07:53:06.411 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout [global] 2026-03-07T07:53:06.411 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout fsid = 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:53:06.411 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-07T07:53:06.411 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.107:3300,v1:192.168.123.107:6789] 2026-03-07T07:53:06.411 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-07T07:53:06.411 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-07T07:53:06.411 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-07T07:53:06.411 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-07T07:53:06.411 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 2026-03-07T07:53:06.411 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-07T07:53:06.411 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-07T07:53:06.411 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 2026-03-07T07:53:06.412 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout [osd] 2026-03-07T07:53:06.412 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-07T07:53:06.412 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-07T07:53:06.412 INFO:teuthology.orchestra.run.vm07.stdout:Enabling cephadm module... 2026-03-07T07:53:07.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:06 vm07 bash[17031]: cluster 2026-03-07T06:53:05.671187+0000 mon.vm07 (mon.0) 32 : cluster [DBG] mgrmap e4: vm07.yrfcuj(active, since 2s) 2026-03-07T07:53:07.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:06 vm07 bash[17031]: cluster 2026-03-07T06:53:05.671187+0000 mon.vm07 (mon.0) 32 : cluster [DBG] mgrmap e4: vm07.yrfcuj(active, since 2s) 2026-03-07T07:53:07.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:06 vm07 bash[17031]: audit 2026-03-07T06:53:06.016513+0000 mon.vm07 (mon.0) 33 : audit [DBG] from='client.? 192.168.123.107:0/757883837' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T07:53:07.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:06 vm07 bash[17031]: audit 2026-03-07T06:53:06.016513+0000 mon.vm07 (mon.0) 33 : audit [DBG] from='client.? 192.168.123.107:0/757883837' entity='client.admin' cmd=[{"prefix": "status", "format": "json-pretty"}]: dispatch 2026-03-07T07:53:07.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:06 vm07 bash[17031]: audit 2026-03-07T06:53:06.366159+0000 mon.vm07 (mon.0) 34 : audit [INF] from='client.? 192.168.123.107:0/689844558' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-07T07:53:07.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:06 vm07 bash[17031]: audit 2026-03-07T06:53:06.366159+0000 mon.vm07 (mon.0) 34 : audit [INF] from='client.? 192.168.123.107:0/689844558' entity='client.admin' cmd=[{"prefix": "config assimilate-conf"}]: dispatch 2026-03-07T07:53:07.960 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:07 vm07 bash[17031]: audit 2026-03-07T06:53:06.743027+0000 mon.vm07 (mon.0) 35 : audit [INF] from='client.? 192.168.123.107:0/3569480277' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-07T07:53:07.960 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:07 vm07 bash[17031]: audit 2026-03-07T06:53:06.743027+0000 mon.vm07 (mon.0) 35 : audit [INF] from='client.? 192.168.123.107:0/3569480277' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "cephadm"}]: dispatch 2026-03-07T07:53:08.170 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout { 2026-03-07T07:53:08.170 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 5, 2026-03-07T07:53:08.170 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-07T07:53:08.170 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "active_name": "vm07.yrfcuj", 2026-03-07T07:53:08.170 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-07T07:53:08.170 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout } 2026-03-07T07:53:08.170 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for the mgr to restart... 2026-03-07T07:53:08.170 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for mgr epoch 5... 2026-03-07T07:53:09.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:08 vm07 bash[17031]: audit 2026-03-07T06:53:07.678034+0000 mon.vm07 (mon.0) 36 : audit [INF] from='client.? 192.168.123.107:0/3569480277' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-07T07:53:09.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:08 vm07 bash[17031]: audit 2026-03-07T06:53:07.678034+0000 mon.vm07 (mon.0) 36 : audit [INF] from='client.? 192.168.123.107:0/3569480277' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-07T07:53:09.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:08 vm07 bash[17031]: cluster 2026-03-07T06:53:07.684208+0000 mon.vm07 (mon.0) 37 : cluster [DBG] mgrmap e5: vm07.yrfcuj(active, since 4s) 2026-03-07T07:53:09.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:08 vm07 bash[17031]: cluster 2026-03-07T06:53:07.684208+0000 mon.vm07 (mon.0) 37 : cluster [DBG] mgrmap e5: vm07.yrfcuj(active, since 4s) 2026-03-07T07:53:09.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:08 vm07 bash[17031]: audit 2026-03-07T06:53:08.107653+0000 mon.vm07 (mon.0) 38 : audit [DBG] from='client.? 192.168.123.107:0/2803343664' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-07T07:53:09.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:08 vm07 bash[17031]: audit 2026-03-07T06:53:08.107653+0000 mon.vm07 (mon.0) 38 : audit [DBG] from='client.? 192.168.123.107:0/2803343664' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: cluster 2026-03-07T06:53:17.025291+0000 mon.vm07 (mon.0) 39 : cluster [INF] Active manager daemon vm07.yrfcuj restarted 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: cluster 2026-03-07T06:53:17.025291+0000 mon.vm07 (mon.0) 39 : cluster [INF] Active manager daemon vm07.yrfcuj restarted 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: cluster 2026-03-07T06:53:17.025491+0000 mon.vm07 (mon.0) 40 : cluster [INF] Activating manager daemon vm07.yrfcuj 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: cluster 2026-03-07T06:53:17.025491+0000 mon.vm07 (mon.0) 40 : cluster [INF] Activating manager daemon vm07.yrfcuj 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: cluster 2026-03-07T06:53:17.031314+0000 mon.vm07 (mon.0) 41 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: cluster 2026-03-07T06:53:17.031314+0000 mon.vm07 (mon.0) 41 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: cluster 2026-03-07T06:53:17.031403+0000 mon.vm07 (mon.0) 42 : cluster [DBG] mgrmap e6: vm07.yrfcuj(active, starting, since 0.00599427s) 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: cluster 2026-03-07T06:53:17.031403+0000 mon.vm07 (mon.0) 42 : cluster [DBG] mgrmap e6: vm07.yrfcuj(active, starting, since 0.00599427s) 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.033295+0000 mon.vm07 (mon.0) 43 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.033295+0000 mon.vm07 (mon.0) 43 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.033390+0000 mon.vm07 (mon.0) 44 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr metadata", "who": "vm07.yrfcuj", "id": "vm07.yrfcuj"}]: dispatch 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.033390+0000 mon.vm07 (mon.0) 44 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr metadata", "who": "vm07.yrfcuj", "id": "vm07.yrfcuj"}]: dispatch 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.035011+0000 mon.vm07 (mon.0) 45 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.035011+0000 mon.vm07 (mon.0) 45 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.035164+0000 mon.vm07 (mon.0) 46 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.035164+0000 mon.vm07 (mon.0) 46 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.035325+0000 mon.vm07 (mon.0) 47 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.035325+0000 mon.vm07 (mon.0) 47 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: cluster 2026-03-07T06:53:17.041033+0000 mon.vm07 (mon.0) 48 : cluster [INF] Manager daemon vm07.yrfcuj is now available 2026-03-07T07:53:17.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: cluster 2026-03-07T06:53:17.041033+0000 mon.vm07 (mon.0) 48 : cluster [INF] Manager daemon vm07.yrfcuj is now available 2026-03-07T07:53:17.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.053659+0000 mon.vm07 (mon.0) 49 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:17.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.053659+0000 mon.vm07 (mon.0) 49 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:17.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.056100+0000 mon.vm07 (mon.0) 50 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:17.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.056100+0000 mon.vm07 (mon.0) 50 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:17.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.071003+0000 mon.vm07 (mon.0) 51 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/mirror_snapshot_schedule"}]: dispatch 2026-03-07T07:53:17.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.071003+0000 mon.vm07 (mon.0) 51 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/mirror_snapshot_schedule"}]: dispatch 2026-03-07T07:53:17.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.071963+0000 mon.vm07 (mon.0) 52 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:53:17.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.071963+0000 mon.vm07 (mon.0) 52 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:53:17.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.074442+0000 mon.vm07 (mon.0) 53 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:53:17.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.074442+0000 mon.vm07 (mon.0) 53 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:53:17.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.076261+0000 mon.vm07 (mon.0) 54 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/trash_purge_schedule"}]: dispatch 2026-03-07T07:53:17.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:17 vm07 bash[17031]: audit 2026-03-07T06:53:17.076261+0000 mon.vm07 (mon.0) 54 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/trash_purge_schedule"}]: dispatch 2026-03-07T07:53:18.100 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout { 2026-03-07T07:53:18.100 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 7, 2026-03-07T07:53:18.100 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-07T07:53:18.100 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout } 2026-03-07T07:53:18.100 INFO:teuthology.orchestra.run.vm07.stdout:mgr epoch 5 is available 2026-03-07T07:53:18.100 INFO:teuthology.orchestra.run.vm07.stdout:Setting orchestrator backend to cephadm... 2026-03-07T07:53:18.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:18 vm07 bash[17031]: cephadm 2026-03-07T06:53:17.051609+0000 mgr.vm07.yrfcuj (mgr.14124) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-07T07:53:18.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:18 vm07 bash[17031]: cephadm 2026-03-07T06:53:17.051609+0000 mgr.vm07.yrfcuj (mgr.14124) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-07T07:53:18.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:18 vm07 bash[17031]: audit 2026-03-07T06:53:17.942644+0000 mon.vm07 (mon.0) 55 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:18.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:18 vm07 bash[17031]: audit 2026-03-07T06:53:17.942644+0000 mon.vm07 (mon.0) 55 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:18.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:18 vm07 bash[17031]: audit 2026-03-07T06:53:17.945949+0000 mon.vm07 (mon.0) 56 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:18.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:18 vm07 bash[17031]: audit 2026-03-07T06:53:17.945949+0000 mon.vm07 (mon.0) 56 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:18.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:18 vm07 bash[17031]: cluster 2026-03-07T06:53:18.034809+0000 mon.vm07 (mon.0) 57 : cluster [DBG] mgrmap e7: vm07.yrfcuj(active, since 1.0094s) 2026-03-07T07:53:18.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:18 vm07 bash[17031]: cluster 2026-03-07T06:53:18.034809+0000 mon.vm07 (mon.0) 57 : cluster [DBG] mgrmap e7: vm07.yrfcuj(active, since 1.0094s) 2026-03-07T07:53:18.872 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-07T07:53:18.872 INFO:teuthology.orchestra.run.vm07.stdout:Generating ssh key... 2026-03-07T07:53:19.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:19 vm07 bash[17031]: audit 2026-03-07T06:53:18.036056+0000 mgr.vm07.yrfcuj (mgr.14124) 2 : audit [DBG] from='client.14128 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-07T07:53:19.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:19 vm07 bash[17031]: audit 2026-03-07T06:53:18.036056+0000 mgr.vm07.yrfcuj (mgr.14124) 2 : audit [DBG] from='client.14128 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-07T07:53:19.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:19 vm07 bash[17031]: audit 2026-03-07T06:53:18.040016+0000 mgr.vm07.yrfcuj (mgr.14124) 3 : audit [DBG] from='client.14128 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-07T07:53:19.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:19 vm07 bash[17031]: audit 2026-03-07T06:53:18.040016+0000 mgr.vm07.yrfcuj (mgr.14124) 3 : audit [DBG] from='client.14128 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-07T07:53:19.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:19 vm07 bash[17031]: audit 2026-03-07T06:53:18.464962+0000 mgr.vm07.yrfcuj (mgr.14124) 4 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:19.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:19 vm07 bash[17031]: audit 2026-03-07T06:53:18.464962+0000 mgr.vm07.yrfcuj (mgr.14124) 4 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:19.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:19 vm07 bash[17031]: audit 2026-03-07T06:53:18.469689+0000 mon.vm07 (mon.0) 58 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:19.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:19 vm07 bash[17031]: audit 2026-03-07T06:53:18.469689+0000 mon.vm07 (mon.0) 58 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:19.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:19 vm07 bash[17031]: audit 2026-03-07T06:53:18.474805+0000 mon.vm07 (mon.0) 59 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:53:19.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:19 vm07 bash[17031]: audit 2026-03-07T06:53:18.474805+0000 mon.vm07 (mon.0) 59 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:53:19.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:19 vm07 bash[17031]: cephadm 2026-03-07T06:53:18.670216+0000 mgr.vm07.yrfcuj (mgr.14124) 5 : cephadm [INF] [07/Mar/2026:06:53:18] ENGINE Bus STARTING 2026-03-07T07:53:19.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:19 vm07 bash[17031]: cephadm 2026-03-07T06:53:18.670216+0000 mgr.vm07.yrfcuj (mgr.14124) 5 : cephadm [INF] [07/Mar/2026:06:53:18] ENGINE Bus STARTING 2026-03-07T07:53:19.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:19 vm07 bash[17031]: audit 2026-03-07T06:53:18.881741+0000 mon.vm07 (mon.0) 60 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:53:19.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:19 vm07 bash[17031]: audit 2026-03-07T06:53:18.881741+0000 mon.vm07 (mon.0) 60 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:53:19.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:19 vm07 bash[17031]: audit 2026-03-07T06:53:19.281266+0000 mon.vm07 (mon.0) 61 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:19.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:19 vm07 bash[17031]: audit 2026-03-07T06:53:19.281266+0000 mon.vm07 (mon.0) 61 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:19.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:19 vm07 bash[17031]: audit 2026-03-07T06:53:19.283773+0000 mon.vm07 (mon.0) 62 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:19.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:19 vm07 bash[17031]: audit 2026-03-07T06:53:19.283773+0000 mon.vm07 (mon.0) 62 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:19.688 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCma9vPL3G8GR2zNAhww92/7xWriUDzymb+ggknjkQJgZQ/o20zmUoedNQsCuk2/+qLNjIcB7ZIbHL5RKJt66OTLPFi45Kjx2W3tuXJLfUn+zzFyGr2eoF9boZcn/er9I/OVMwD2YyNgAjaYFjWY3rMqdCdcrJoN1TezIX6sqZHcvIggkwV+SD0DgQrS6KlvbP1QuoaCxySiUp737p3axabZftIlLL79MiJOjhcSjXj7PRh2Stkc+TC2XPT7Soj9InkxuYURdWKpy77sfB9qRnULCSCBeP5GRg7Aclqks8hapLghoPg5JL9VjwQmF7rgduQd8HiCq4DcIKU+kPZVJSJluwn9ugWlJlM2XlF6brEIo2qqQDNTtdiWrTP8n4HEVCtIC0BRCcVd22ft50dDkZyqJXUnT6CLevS9ZOrozroKZ1+W8cerjS40SYqnMCdA1bZXfKZMjCkSxMwWJta7SV5Xh4jcJUSZkRnSO0HakNL8QtTa1id1LYeZgq6y9uACAE= ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:53:19.688 INFO:teuthology.orchestra.run.vm07.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-07T07:53:19.688 INFO:teuthology.orchestra.run.vm07.stdout:Adding key to root@localhost authorized_keys... 2026-03-07T07:53:19.688 INFO:teuthology.orchestra.run.vm07.stdout:Adding host vm07... 2026-03-07T07:53:20.771 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:20 vm07 bash[17031]: cephadm 2026-03-07T06:53:18.779831+0000 mgr.vm07.yrfcuj (mgr.14124) 6 : cephadm [INF] [07/Mar/2026:06:53:18] ENGINE Serving on https://192.168.123.107:7150 2026-03-07T07:53:20.771 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:20 vm07 bash[17031]: cephadm 2026-03-07T06:53:18.779831+0000 mgr.vm07.yrfcuj (mgr.14124) 6 : cephadm [INF] [07/Mar/2026:06:53:18] ENGINE Serving on https://192.168.123.107:7150 2026-03-07T07:53:20.771 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:20 vm07 bash[17031]: cephadm 2026-03-07T06:53:18.780382+0000 mgr.vm07.yrfcuj (mgr.14124) 7 : cephadm [INF] [07/Mar/2026:06:53:18] ENGINE Client ('192.168.123.107', 37048) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-07T07:53:20.772 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:20 vm07 bash[17031]: cephadm 2026-03-07T06:53:18.780382+0000 mgr.vm07.yrfcuj (mgr.14124) 7 : cephadm [INF] [07/Mar/2026:06:53:18] ENGINE Client ('192.168.123.107', 37048) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-07T07:53:20.772 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:20 vm07 bash[17031]: audit 2026-03-07T06:53:18.826825+0000 mgr.vm07.yrfcuj (mgr.14124) 8 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:20.772 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:20 vm07 bash[17031]: audit 2026-03-07T06:53:18.826825+0000 mgr.vm07.yrfcuj (mgr.14124) 8 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:20.772 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:20 vm07 bash[17031]: cephadm 2026-03-07T06:53:18.881348+0000 mgr.vm07.yrfcuj (mgr.14124) 9 : cephadm [INF] [07/Mar/2026:06:53:18] ENGINE Serving on http://192.168.123.107:8765 2026-03-07T07:53:20.772 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:20 vm07 bash[17031]: cephadm 2026-03-07T06:53:18.881348+0000 mgr.vm07.yrfcuj (mgr.14124) 9 : cephadm [INF] [07/Mar/2026:06:53:18] ENGINE Serving on http://192.168.123.107:8765 2026-03-07T07:53:20.772 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:20 vm07 bash[17031]: cephadm 2026-03-07T06:53:18.881388+0000 mgr.vm07.yrfcuj (mgr.14124) 10 : cephadm [INF] [07/Mar/2026:06:53:18] ENGINE Bus STARTED 2026-03-07T07:53:20.772 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:20 vm07 bash[17031]: cephadm 2026-03-07T06:53:18.881388+0000 mgr.vm07.yrfcuj (mgr.14124) 10 : cephadm [INF] [07/Mar/2026:06:53:18] ENGINE Bus STARTED 2026-03-07T07:53:20.772 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:20 vm07 bash[17031]: audit 2026-03-07T06:53:19.179532+0000 mgr.vm07.yrfcuj (mgr.14124) 11 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:20.772 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:20 vm07 bash[17031]: audit 2026-03-07T06:53:19.179532+0000 mgr.vm07.yrfcuj (mgr.14124) 11 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:20.772 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:20 vm07 bash[17031]: cephadm 2026-03-07T06:53:19.179747+0000 mgr.vm07.yrfcuj (mgr.14124) 12 : cephadm [INF] Generating ssh key... 2026-03-07T07:53:20.772 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:20 vm07 bash[17031]: cephadm 2026-03-07T06:53:19.179747+0000 mgr.vm07.yrfcuj (mgr.14124) 12 : cephadm [INF] Generating ssh key... 2026-03-07T07:53:20.772 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:20 vm07 bash[17031]: cluster 2026-03-07T06:53:19.479108+0000 mon.vm07 (mon.0) 63 : cluster [DBG] mgrmap e8: vm07.yrfcuj(active, since 2s) 2026-03-07T07:53:20.772 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:20 vm07 bash[17031]: cluster 2026-03-07T06:53:19.479108+0000 mon.vm07 (mon.0) 63 : cluster [DBG] mgrmap e8: vm07.yrfcuj(active, since 2s) 2026-03-07T07:53:20.772 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:20 vm07 bash[17031]: audit 2026-03-07T06:53:19.638419+0000 mgr.vm07.yrfcuj (mgr.14124) 13 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:20.772 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:20 vm07 bash[17031]: audit 2026-03-07T06:53:19.638419+0000 mgr.vm07.yrfcuj (mgr.14124) 13 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:21.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:21 vm07 bash[17031]: audit 2026-03-07T06:53:19.981772+0000 mgr.vm07.yrfcuj (mgr.14124) 14 : audit [DBG] from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm07", "addr": "192.168.123.107", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:21.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:21 vm07 bash[17031]: audit 2026-03-07T06:53:19.981772+0000 mgr.vm07.yrfcuj (mgr.14124) 14 : audit [DBG] from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm07", "addr": "192.168.123.107", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:22.669 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout Added host 'vm07' with addr '192.168.123.107' 2026-03-07T07:53:22.669 INFO:teuthology.orchestra.run.vm07.stdout:Deploying mon service with default placement... 2026-03-07T07:53:22.740 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:22 vm07 bash[17031]: cephadm 2026-03-07T06:53:20.859113+0000 mgr.vm07.yrfcuj (mgr.14124) 15 : cephadm [INF] Deploying cephadm binary to vm07 2026-03-07T07:53:22.740 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:22 vm07 bash[17031]: cephadm 2026-03-07T06:53:20.859113+0000 mgr.vm07.yrfcuj (mgr.14124) 15 : cephadm [INF] Deploying cephadm binary to vm07 2026-03-07T07:53:23.081 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-07T07:53:23.081 INFO:teuthology.orchestra.run.vm07.stdout:Deploying mgr service with default placement... 2026-03-07T07:53:23.473 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-07T07:53:23.473 INFO:teuthology.orchestra.run.vm07.stdout:Deploying crash service with default placement... 2026-03-07T07:53:23.672 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:23 vm07 bash[17031]: audit 2026-03-07T06:53:22.590673+0000 mon.vm07 (mon.0) 64 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:23.672 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:23 vm07 bash[17031]: audit 2026-03-07T06:53:22.590673+0000 mon.vm07 (mon.0) 64 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:23.672 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:23 vm07 bash[17031]: cephadm 2026-03-07T06:53:22.591107+0000 mgr.vm07.yrfcuj (mgr.14124) 16 : cephadm [INF] Added host vm07 2026-03-07T07:53:23.672 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:23 vm07 bash[17031]: cephadm 2026-03-07T06:53:22.591107+0000 mgr.vm07.yrfcuj (mgr.14124) 16 : cephadm [INF] Added host vm07 2026-03-07T07:53:23.672 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:23 vm07 bash[17031]: audit 2026-03-07T06:53:22.591538+0000 mon.vm07 (mon.0) 65 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:53:23.672 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:23 vm07 bash[17031]: audit 2026-03-07T06:53:22.591538+0000 mon.vm07 (mon.0) 65 : audit [DBG] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:53:23.672 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:23 vm07 bash[17031]: audit 2026-03-07T06:53:23.033423+0000 mon.vm07 (mon.0) 66 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:23.672 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:23 vm07 bash[17031]: audit 2026-03-07T06:53:23.033423+0000 mon.vm07 (mon.0) 66 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:23.672 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:23 vm07 bash[17031]: audit 2026-03-07T06:53:23.402078+0000 mon.vm07 (mon.0) 67 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:23.672 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:23 vm07 bash[17031]: audit 2026-03-07T06:53:23.402078+0000 mon.vm07 (mon.0) 67 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:23.895 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout Scheduled crash update... 2026-03-07T07:53:23.895 INFO:teuthology.orchestra.run.vm07.stdout:Deploying ceph-exporter service with default placement... 2026-03-07T07:53:24.315 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout Scheduled ceph-exporter update... 2026-03-07T07:53:24.315 INFO:teuthology.orchestra.run.vm07.stdout:Deploying prometheus service with default placement... 2026-03-07T07:53:24.673 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout Scheduled prometheus update... 2026-03-07T07:53:24.673 INFO:teuthology.orchestra.run.vm07.stdout:Deploying grafana service with default placement... 2026-03-07T07:53:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:24 vm07 bash[17031]: audit 2026-03-07T06:53:23.029859+0000 mgr.vm07.yrfcuj (mgr.14124) 17 : audit [DBG] from='client.14146 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:24 vm07 bash[17031]: audit 2026-03-07T06:53:23.029859+0000 mgr.vm07.yrfcuj (mgr.14124) 17 : audit [DBG] from='client.14146 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:24 vm07 bash[17031]: cephadm 2026-03-07T06:53:23.030675+0000 mgr.vm07.yrfcuj (mgr.14124) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-07T07:53:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:24 vm07 bash[17031]: cephadm 2026-03-07T06:53:23.030675+0000 mgr.vm07.yrfcuj (mgr.14124) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-07T07:53:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:24 vm07 bash[17031]: audit 2026-03-07T06:53:23.395989+0000 mgr.vm07.yrfcuj (mgr.14124) 19 : audit [DBG] from='client.14148 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:24 vm07 bash[17031]: audit 2026-03-07T06:53:23.395989+0000 mgr.vm07.yrfcuj (mgr.14124) 19 : audit [DBG] from='client.14148 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:24 vm07 bash[17031]: cephadm 2026-03-07T06:53:23.396860+0000 mgr.vm07.yrfcuj (mgr.14124) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-07T07:53:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:24 vm07 bash[17031]: cephadm 2026-03-07T06:53:23.396860+0000 mgr.vm07.yrfcuj (mgr.14124) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-07T07:53:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:24 vm07 bash[17031]: audit 2026-03-07T06:53:23.710543+0000 mon.vm07 (mon.0) 68 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:24 vm07 bash[17031]: audit 2026-03-07T06:53:23.710543+0000 mon.vm07 (mon.0) 68 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:24 vm07 bash[17031]: audit 2026-03-07T06:53:23.840890+0000 mon.vm07 (mon.0) 69 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:24 vm07 bash[17031]: audit 2026-03-07T06:53:23.840890+0000 mon.vm07 (mon.0) 69 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:24 vm07 bash[17031]: audit 2026-03-07T06:53:24.034105+0000 mon.vm07 (mon.0) 70 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:24 vm07 bash[17031]: audit 2026-03-07T06:53:24.034105+0000 mon.vm07 (mon.0) 70 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:24 vm07 bash[17031]: audit 2026-03-07T06:53:24.253281+0000 mon.vm07 (mon.0) 71 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:24 vm07 bash[17031]: audit 2026-03-07T06:53:24.253281+0000 mon.vm07 (mon.0) 71 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:24 vm07 bash[17031]: audit 2026-03-07T06:53:24.613072+0000 mon.vm07 (mon.0) 72 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:24 vm07 bash[17031]: audit 2026-03-07T06:53:24.613072+0000 mon.vm07 (mon.0) 72 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:25.038 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout Scheduled grafana update... 2026-03-07T07:53:25.038 INFO:teuthology.orchestra.run.vm07.stdout:Deploying node-exporter service with default placement... 2026-03-07T07:53:25.379 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout Scheduled node-exporter update... 2026-03-07T07:53:25.379 INFO:teuthology.orchestra.run.vm07.stdout:Deploying alertmanager service with default placement... 2026-03-07T07:53:25.745 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout Scheduled alertmanager update... 2026-03-07T07:53:26.058 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:25 vm07 bash[17031]: audit 2026-03-07T06:53:23.836878+0000 mgr.vm07.yrfcuj (mgr.14124) 21 : audit [DBG] from='client.14150 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "crash", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:26.058 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:25 vm07 bash[17031]: audit 2026-03-07T06:53:23.836878+0000 mgr.vm07.yrfcuj (mgr.14124) 21 : audit [DBG] from='client.14150 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "crash", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:26.058 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:25 vm07 bash[17031]: cephadm 2026-03-07T06:53:23.837557+0000 mgr.vm07.yrfcuj (mgr.14124) 22 : cephadm [INF] Saving service crash spec with placement * 2026-03-07T07:53:26.058 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:25 vm07 bash[17031]: cephadm 2026-03-07T06:53:23.837557+0000 mgr.vm07.yrfcuj (mgr.14124) 22 : cephadm [INF] Saving service crash spec with placement * 2026-03-07T07:53:26.058 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:25 vm07 bash[17031]: audit 2026-03-07T06:53:24.246786+0000 mgr.vm07.yrfcuj (mgr.14124) 23 : audit [DBG] from='client.14152 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "ceph-exporter", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:26.058 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:25 vm07 bash[17031]: audit 2026-03-07T06:53:24.246786+0000 mgr.vm07.yrfcuj (mgr.14124) 23 : audit [DBG] from='client.14152 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "ceph-exporter", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:26.058 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:25 vm07 bash[17031]: cephadm 2026-03-07T06:53:24.247567+0000 mgr.vm07.yrfcuj (mgr.14124) 24 : cephadm [INF] Saving service ceph-exporter spec with placement * 2026-03-07T07:53:26.058 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:25 vm07 bash[17031]: cephadm 2026-03-07T06:53:24.247567+0000 mgr.vm07.yrfcuj (mgr.14124) 24 : cephadm [INF] Saving service ceph-exporter spec with placement * 2026-03-07T07:53:26.058 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:25 vm07 bash[17031]: audit 2026-03-07T06:53:24.609416+0000 mgr.vm07.yrfcuj (mgr.14124) 25 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "prometheus", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:26.058 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:25 vm07 bash[17031]: audit 2026-03-07T06:53:24.609416+0000 mgr.vm07.yrfcuj (mgr.14124) 25 : audit [DBG] from='client.14154 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "prometheus", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:26.058 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:25 vm07 bash[17031]: cephadm 2026-03-07T06:53:24.610250+0000 mgr.vm07.yrfcuj (mgr.14124) 26 : cephadm [INF] Saving service prometheus spec with placement count:1 2026-03-07T07:53:26.058 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:25 vm07 bash[17031]: cephadm 2026-03-07T06:53:24.610250+0000 mgr.vm07.yrfcuj (mgr.14124) 26 : cephadm [INF] Saving service prometheus spec with placement count:1 2026-03-07T07:53:26.058 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:25 vm07 bash[17031]: audit 2026-03-07T06:53:24.990782+0000 mon.vm07 (mon.0) 73 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:26.058 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:25 vm07 bash[17031]: audit 2026-03-07T06:53:24.990782+0000 mon.vm07 (mon.0) 73 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:26.058 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:25 vm07 bash[17031]: audit 2026-03-07T06:53:25.335898+0000 mon.vm07 (mon.0) 74 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:26.058 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:25 vm07 bash[17031]: audit 2026-03-07T06:53:25.335898+0000 mon.vm07 (mon.0) 74 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:26.058 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:25 vm07 bash[17031]: audit 2026-03-07T06:53:25.698469+0000 mon.vm07 (mon.0) 75 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:26.058 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:25 vm07 bash[17031]: audit 2026-03-07T06:53:25.698469+0000 mon.vm07 (mon.0) 75 : audit [INF] from='mgr.14124 192.168.123.107:0/359001630' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:26.440 INFO:teuthology.orchestra.run.vm07.stdout:Enabling the dashboard module... 2026-03-07T07:53:27.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:26 vm07 bash[17031]: audit 2026-03-07T06:53:24.987178+0000 mgr.vm07.yrfcuj (mgr.14124) 27 : audit [DBG] from='client.14156 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "grafana", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:27.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:26 vm07 bash[17031]: audit 2026-03-07T06:53:24.987178+0000 mgr.vm07.yrfcuj (mgr.14124) 27 : audit [DBG] from='client.14156 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "grafana", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:27.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:26 vm07 bash[17031]: cephadm 2026-03-07T06:53:24.987947+0000 mgr.vm07.yrfcuj (mgr.14124) 28 : cephadm [INF] Saving service grafana spec with placement count:1 2026-03-07T07:53:27.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:26 vm07 bash[17031]: cephadm 2026-03-07T06:53:24.987947+0000 mgr.vm07.yrfcuj (mgr.14124) 28 : cephadm [INF] Saving service grafana spec with placement count:1 2026-03-07T07:53:27.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:26 vm07 bash[17031]: audit 2026-03-07T06:53:25.332738+0000 mgr.vm07.yrfcuj (mgr.14124) 29 : audit [DBG] from='client.14158 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "node-exporter", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:27.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:26 vm07 bash[17031]: audit 2026-03-07T06:53:25.332738+0000 mgr.vm07.yrfcuj (mgr.14124) 29 : audit [DBG] from='client.14158 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "node-exporter", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:27.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:26 vm07 bash[17031]: cephadm 2026-03-07T06:53:25.333413+0000 mgr.vm07.yrfcuj (mgr.14124) 30 : cephadm [INF] Saving service node-exporter spec with placement * 2026-03-07T07:53:27.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:26 vm07 bash[17031]: cephadm 2026-03-07T06:53:25.333413+0000 mgr.vm07.yrfcuj (mgr.14124) 30 : cephadm [INF] Saving service node-exporter spec with placement * 2026-03-07T07:53:27.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:26 vm07 bash[17031]: audit 2026-03-07T06:53:25.695053+0000 mgr.vm07.yrfcuj (mgr.14124) 31 : audit [DBG] from='client.14160 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "alertmanager", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:27.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:26 vm07 bash[17031]: audit 2026-03-07T06:53:25.695053+0000 mgr.vm07.yrfcuj (mgr.14124) 31 : audit [DBG] from='client.14160 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "alertmanager", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:27.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:26 vm07 bash[17031]: cephadm 2026-03-07T06:53:25.695715+0000 mgr.vm07.yrfcuj (mgr.14124) 32 : cephadm [INF] Saving service alertmanager spec with placement count:1 2026-03-07T07:53:27.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:26 vm07 bash[17031]: cephadm 2026-03-07T06:53:25.695715+0000 mgr.vm07.yrfcuj (mgr.14124) 32 : cephadm [INF] Saving service alertmanager spec with placement count:1 2026-03-07T07:53:27.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:26 vm07 bash[17031]: audit 2026-03-07T06:53:26.039839+0000 mon.vm07 (mon.0) 76 : audit [INF] from='client.? 192.168.123.107:0/1059279599' entity='client.admin' 2026-03-07T07:53:27.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:26 vm07 bash[17031]: audit 2026-03-07T06:53:26.039839+0000 mon.vm07 (mon.0) 76 : audit [INF] from='client.? 192.168.123.107:0/1059279599' entity='client.admin' 2026-03-07T07:53:27.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:26 vm07 bash[17031]: audit 2026-03-07T06:53:26.392000+0000 mon.vm07 (mon.0) 77 : audit [INF] from='client.? 192.168.123.107:0/272525019' entity='client.admin' 2026-03-07T07:53:27.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:26 vm07 bash[17031]: audit 2026-03-07T06:53:26.392000+0000 mon.vm07 (mon.0) 77 : audit [INF] from='client.? 192.168.123.107:0/272525019' entity='client.admin' 2026-03-07T07:53:28.015 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:27 vm07 bash[17031]: audit 2026-03-07T06:53:26.816274+0000 mon.vm07 (mon.0) 78 : audit [INF] from='client.? 192.168.123.107:0/4038423287' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-07T07:53:28.015 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:27 vm07 bash[17031]: audit 2026-03-07T06:53:26.816274+0000 mon.vm07 (mon.0) 78 : audit [INF] from='client.? 192.168.123.107:0/4038423287' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-07T07:53:28.256 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout { 2026-03-07T07:53:28.256 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "epoch": 9, 2026-03-07T07:53:28.256 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-07T07:53:28.256 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "active_name": "vm07.yrfcuj", 2026-03-07T07:53:28.256 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-07T07:53:28.256 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout } 2026-03-07T07:53:28.256 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for the mgr to restart... 2026-03-07T07:53:28.256 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for mgr epoch 9... 2026-03-07T07:53:29.065 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:28 vm07 bash[17031]: audit 2026-03-07T06:53:27.719249+0000 mon.vm07 (mon.0) 79 : audit [INF] from='client.? 192.168.123.107:0/4038423287' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-07T07:53:29.065 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:28 vm07 bash[17031]: audit 2026-03-07T06:53:27.719249+0000 mon.vm07 (mon.0) 79 : audit [INF] from='client.? 192.168.123.107:0/4038423287' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-07T07:53:29.065 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:28 vm07 bash[17031]: cluster 2026-03-07T06:53:27.721510+0000 mon.vm07 (mon.0) 80 : cluster [DBG] mgrmap e9: vm07.yrfcuj(active, since 10s) 2026-03-07T07:53:29.065 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:28 vm07 bash[17031]: cluster 2026-03-07T06:53:27.721510+0000 mon.vm07 (mon.0) 80 : cluster [DBG] mgrmap e9: vm07.yrfcuj(active, since 10s) 2026-03-07T07:53:29.065 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:28 vm07 bash[17031]: audit 2026-03-07T06:53:28.187697+0000 mon.vm07 (mon.0) 81 : audit [DBG] from='client.? 192.168.123.107:0/2783722348' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-07T07:53:29.065 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:28 vm07 bash[17031]: audit 2026-03-07T06:53:28.187697+0000 mon.vm07 (mon.0) 81 : audit [DBG] from='client.? 192.168.123.107:0/2783722348' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: cluster 2026-03-07T06:53:37.191080+0000 mon.vm07 (mon.0) 82 : cluster [INF] Active manager daemon vm07.yrfcuj restarted 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: cluster 2026-03-07T06:53:37.191080+0000 mon.vm07 (mon.0) 82 : cluster [INF] Active manager daemon vm07.yrfcuj restarted 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: cluster 2026-03-07T06:53:37.191299+0000 mon.vm07 (mon.0) 83 : cluster [INF] Activating manager daemon vm07.yrfcuj 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: cluster 2026-03-07T06:53:37.191299+0000 mon.vm07 (mon.0) 83 : cluster [INF] Activating manager daemon vm07.yrfcuj 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: cluster 2026-03-07T06:53:37.197363+0000 mon.vm07 (mon.0) 84 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: cluster 2026-03-07T06:53:37.197363+0000 mon.vm07 (mon.0) 84 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: cluster 2026-03-07T06:53:37.197482+0000 mon.vm07 (mon.0) 85 : cluster [DBG] mgrmap e10: vm07.yrfcuj(active, starting, since 0.0062798s) 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: cluster 2026-03-07T06:53:37.197482+0000 mon.vm07 (mon.0) 85 : cluster [DBG] mgrmap e10: vm07.yrfcuj(active, starting, since 0.0062798s) 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: audit 2026-03-07T06:53:37.199545+0000 mon.vm07 (mon.0) 86 : audit [DBG] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: audit 2026-03-07T06:53:37.199545+0000 mon.vm07 (mon.0) 86 : audit [DBG] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: audit 2026-03-07T06:53:37.200453+0000 mon.vm07 (mon.0) 87 : audit [DBG] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr metadata", "who": "vm07.yrfcuj", "id": "vm07.yrfcuj"}]: dispatch 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: audit 2026-03-07T06:53:37.200453+0000 mon.vm07 (mon.0) 87 : audit [DBG] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr metadata", "who": "vm07.yrfcuj", "id": "vm07.yrfcuj"}]: dispatch 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: audit 2026-03-07T06:53:37.201386+0000 mon.vm07 (mon.0) 88 : audit [DBG] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: audit 2026-03-07T06:53:37.201386+0000 mon.vm07 (mon.0) 88 : audit [DBG] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: audit 2026-03-07T06:53:37.201526+0000 mon.vm07 (mon.0) 89 : audit [DBG] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: audit 2026-03-07T06:53:37.201526+0000 mon.vm07 (mon.0) 89 : audit [DBG] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: audit 2026-03-07T06:53:37.201679+0000 mon.vm07 (mon.0) 90 : audit [DBG] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: audit 2026-03-07T06:53:37.201679+0000 mon.vm07 (mon.0) 90 : audit [DBG] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: cluster 2026-03-07T06:53:37.207002+0000 mon.vm07 (mon.0) 91 : cluster [INF] Manager daemon vm07.yrfcuj is now available 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: cluster 2026-03-07T06:53:37.207002+0000 mon.vm07 (mon.0) 91 : cluster [INF] Manager daemon vm07.yrfcuj is now available 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: audit 2026-03-07T06:53:37.231629+0000 mon.vm07 (mon.0) 92 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/mirror_snapshot_schedule"}]: dispatch 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: audit 2026-03-07T06:53:37.231629+0000 mon.vm07 (mon.0) 92 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/mirror_snapshot_schedule"}]: dispatch 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: audit 2026-03-07T06:53:37.231875+0000 mon.vm07 (mon.0) 93 : audit [DBG] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: audit 2026-03-07T06:53:37.231875+0000 mon.vm07 (mon.0) 93 : audit [DBG] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:53:37.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: audit 2026-03-07T06:53:37.240599+0000 mon.vm07 (mon.0) 94 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/trash_purge_schedule"}]: dispatch 2026-03-07T07:53:37.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:37 vm07 bash[17031]: audit 2026-03-07T06:53:37.240599+0000 mon.vm07 (mon.0) 94 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/trash_purge_schedule"}]: dispatch 2026-03-07T07:53:38.283 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout { 2026-03-07T07:53:38.283 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 11, 2026-03-07T07:53:38.283 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-07T07:53:38.283 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout } 2026-03-07T07:53:38.283 INFO:teuthology.orchestra.run.vm07.stdout:mgr epoch 9 is available 2026-03-07T07:53:38.283 INFO:teuthology.orchestra.run.vm07.stdout:Generating a dashboard self-signed certificate... 2026-03-07T07:53:38.715 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-07T07:53:38.715 INFO:teuthology.orchestra.run.vm07.stdout:Creating initial admin user... 2026-03-07T07:53:39.234 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$E0f3BxKTOPlnzOp5118SGugotppKD.EwsqkPvpmsQwVImQeQVdmPS", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1772866419, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-07T07:53:39.234 INFO:teuthology.orchestra.run.vm07.stdout:Fetching dashboard port number... 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: cluster 2026-03-07T06:53:38.201868+0000 mon.vm07 (mon.0) 95 : cluster [DBG] mgrmap e11: vm07.yrfcuj(active, since 1.01067s) 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: cluster 2026-03-07T06:53:38.201868+0000 mon.vm07 (mon.0) 95 : cluster [DBG] mgrmap e11: vm07.yrfcuj(active, since 1.01067s) 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: cephadm 2026-03-07T06:53:38.243048+0000 mgr.vm07.yrfcuj (mgr.14168) 3 : cephadm [INF] [07/Mar/2026:06:53:38] ENGINE Bus STARTING 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: cephadm 2026-03-07T06:53:38.243048+0000 mgr.vm07.yrfcuj (mgr.14168) 3 : cephadm [INF] [07/Mar/2026:06:53:38] ENGINE Bus STARTING 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: cephadm 2026-03-07T06:53:38.356833+0000 mgr.vm07.yrfcuj (mgr.14168) 4 : cephadm [INF] [07/Mar/2026:06:53:38] ENGINE Serving on https://192.168.123.107:7150 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: cephadm 2026-03-07T06:53:38.356833+0000 mgr.vm07.yrfcuj (mgr.14168) 4 : cephadm [INF] [07/Mar/2026:06:53:38] ENGINE Serving on https://192.168.123.107:7150 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: cephadm 2026-03-07T06:53:38.357666+0000 mgr.vm07.yrfcuj (mgr.14168) 5 : cephadm [INF] [07/Mar/2026:06:53:38] ENGINE Client ('192.168.123.107', 57614) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: cephadm 2026-03-07T06:53:38.357666+0000 mgr.vm07.yrfcuj (mgr.14168) 5 : cephadm [INF] [07/Mar/2026:06:53:38] ENGINE Client ('192.168.123.107', 57614) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: cephadm 2026-03-07T06:53:38.457793+0000 mgr.vm07.yrfcuj (mgr.14168) 6 : cephadm [INF] [07/Mar/2026:06:53:38] ENGINE Serving on http://192.168.123.107:8765 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: cephadm 2026-03-07T06:53:38.457793+0000 mgr.vm07.yrfcuj (mgr.14168) 6 : cephadm [INF] [07/Mar/2026:06:53:38] ENGINE Serving on http://192.168.123.107:8765 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: cephadm 2026-03-07T06:53:38.457853+0000 mgr.vm07.yrfcuj (mgr.14168) 7 : cephadm [INF] [07/Mar/2026:06:53:38] ENGINE Bus STARTED 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: cephadm 2026-03-07T06:53:38.457853+0000 mgr.vm07.yrfcuj (mgr.14168) 7 : cephadm [INF] [07/Mar/2026:06:53:38] ENGINE Bus STARTED 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: audit 2026-03-07T06:53:38.584565+0000 mgr.vm07.yrfcuj (mgr.14168) 8 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: audit 2026-03-07T06:53:38.584565+0000 mgr.vm07.yrfcuj (mgr.14168) 8 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: audit 2026-03-07T06:53:38.667136+0000 mon.vm07 (mon.0) 96 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: audit 2026-03-07T06:53:38.667136+0000 mon.vm07 (mon.0) 96 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: audit 2026-03-07T06:53:38.670632+0000 mon.vm07 (mon.0) 97 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: audit 2026-03-07T06:53:38.670632+0000 mon.vm07 (mon.0) 97 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: audit 2026-03-07T06:53:39.182162+0000 mon.vm07 (mon.0) 98 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:39.540 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:39 vm07 bash[17031]: audit 2026-03-07T06:53:39.182162+0000 mon.vm07 (mon.0) 98 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:39.574 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stdout 8443 2026-03-07T07:53:39.574 INFO:teuthology.orchestra.run.vm07.stdout:firewalld does not appear to be present 2026-03-07T07:53:39.574 INFO:teuthology.orchestra.run.vm07.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-07T07:53:39.575 INFO:teuthology.orchestra.run.vm07.stdout:Ceph Dashboard is now available at: 2026-03-07T07:53:39.575 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:53:39.575 INFO:teuthology.orchestra.run.vm07.stdout: URL: https://vm07.local:8443/ 2026-03-07T07:53:39.575 INFO:teuthology.orchestra.run.vm07.stdout: User: admin 2026-03-07T07:53:39.575 INFO:teuthology.orchestra.run.vm07.stdout: Password: zcwi6jvzmf 2026-03-07T07:53:39.575 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:53:39.575 INFO:teuthology.orchestra.run.vm07.stdout:Saving cluster configuration to /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config directory 2026-03-07T07:53:39.971 INFO:teuthology.orchestra.run.vm07.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-07T07:53:39.972 INFO:teuthology.orchestra.run.vm07.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-07T07:53:39.972 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:53:39.972 INFO:teuthology.orchestra.run.vm07.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-07T07:53:39.972 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:53:39.972 INFO:teuthology.orchestra.run.vm07.stdout:Or, if you are only running a single cluster on this host: 2026-03-07T07:53:39.972 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:53:39.972 INFO:teuthology.orchestra.run.vm07.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-07T07:53:39.972 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:53:39.972 INFO:teuthology.orchestra.run.vm07.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-07T07:53:39.972 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:53:39.972 INFO:teuthology.orchestra.run.vm07.stdout: ceph telemetry on 2026-03-07T07:53:39.972 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:53:39.972 INFO:teuthology.orchestra.run.vm07.stdout:For more information see: 2026-03-07T07:53:39.972 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:53:39.972 INFO:teuthology.orchestra.run.vm07.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-07T07:53:39.972 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:53:39.975 INFO:teuthology.orchestra.run.vm07.stdout:Bootstrap complete. 2026-03-07T07:53:39.992 INFO:tasks.cephadm:Fetching config... 2026-03-07T07:53:39.992 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-07T07:53:39.992 DEBUG:teuthology.orchestra.run.vm07:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-07T07:53:39.995 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-07T07:53:39.995 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-07T07:53:39.995 DEBUG:teuthology.orchestra.run.vm07:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-07T07:53:40.038 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-07T07:53:40.038 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-07T07:53:40.038 DEBUG:teuthology.orchestra.run.vm07:> sudo dd if=/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/keyring of=/dev/stdout 2026-03-07T07:53:40.086 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-07T07:53:40.087 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-07T07:53:40.087 DEBUG:teuthology.orchestra.run.vm07:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-07T07:53:40.130 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-07T07:53:40.131 DEBUG:teuthology.orchestra.run.vm07:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCma9vPL3G8GR2zNAhww92/7xWriUDzymb+ggknjkQJgZQ/o20zmUoedNQsCuk2/+qLNjIcB7ZIbHL5RKJt66OTLPFi45Kjx2W3tuXJLfUn+zzFyGr2eoF9boZcn/er9I/OVMwD2YyNgAjaYFjWY3rMqdCdcrJoN1TezIX6sqZHcvIggkwV+SD0DgQrS6KlvbP1QuoaCxySiUp737p3axabZftIlLL79MiJOjhcSjXj7PRh2Stkc+TC2XPT7Soj9InkxuYURdWKpy77sfB9qRnULCSCBeP5GRg7Aclqks8hapLghoPg5JL9VjwQmF7rgduQd8HiCq4DcIKU+kPZVJSJluwn9ugWlJlM2XlF6brEIo2qqQDNTtdiWrTP8n4HEVCtIC0BRCcVd22ft50dDkZyqJXUnT6CLevS9ZOrozroKZ1+W8cerjS40SYqnMCdA1bZXfKZMjCkSxMwWJta7SV5Xh4jcJUSZkRnSO0HakNL8QtTa1id1LYeZgq6y9uACAE= ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-07T07:53:40.183 INFO:teuthology.orchestra.run.vm07.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCma9vPL3G8GR2zNAhww92/7xWriUDzymb+ggknjkQJgZQ/o20zmUoedNQsCuk2/+qLNjIcB7ZIbHL5RKJt66OTLPFi45Kjx2W3tuXJLfUn+zzFyGr2eoF9boZcn/er9I/OVMwD2YyNgAjaYFjWY3rMqdCdcrJoN1TezIX6sqZHcvIggkwV+SD0DgQrS6KlvbP1QuoaCxySiUp737p3axabZftIlLL79MiJOjhcSjXj7PRh2Stkc+TC2XPT7Soj9InkxuYURdWKpy77sfB9qRnULCSCBeP5GRg7Aclqks8hapLghoPg5JL9VjwQmF7rgduQd8HiCq4DcIKU+kPZVJSJluwn9ugWlJlM2XlF6brEIo2qqQDNTtdiWrTP8n4HEVCtIC0BRCcVd22ft50dDkZyqJXUnT6CLevS9ZOrozroKZ1+W8cerjS40SYqnMCdA1bZXfKZMjCkSxMwWJta7SV5Xh4jcJUSZkRnSO0HakNL8QtTa1id1LYeZgq6y9uACAE= ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:53:40.189 DEBUG:teuthology.orchestra.run.vm09:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCma9vPL3G8GR2zNAhww92/7xWriUDzymb+ggknjkQJgZQ/o20zmUoedNQsCuk2/+qLNjIcB7ZIbHL5RKJt66OTLPFi45Kjx2W3tuXJLfUn+zzFyGr2eoF9boZcn/er9I/OVMwD2YyNgAjaYFjWY3rMqdCdcrJoN1TezIX6sqZHcvIggkwV+SD0DgQrS6KlvbP1QuoaCxySiUp737p3axabZftIlLL79MiJOjhcSjXj7PRh2Stkc+TC2XPT7Soj9InkxuYURdWKpy77sfB9qRnULCSCBeP5GRg7Aclqks8hapLghoPg5JL9VjwQmF7rgduQd8HiCq4DcIKU+kPZVJSJluwn9ugWlJlM2XlF6brEIo2qqQDNTtdiWrTP8n4HEVCtIC0BRCcVd22ft50dDkZyqJXUnT6CLevS9ZOrozroKZ1+W8cerjS40SYqnMCdA1bZXfKZMjCkSxMwWJta7SV5Xh4jcJUSZkRnSO0HakNL8QtTa1id1LYeZgq6y9uACAE= ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-07T07:53:40.200 INFO:teuthology.orchestra.run.vm09.stdout:ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABgQCma9vPL3G8GR2zNAhww92/7xWriUDzymb+ggknjkQJgZQ/o20zmUoedNQsCuk2/+qLNjIcB7ZIbHL5RKJt66OTLPFi45Kjx2W3tuXJLfUn+zzFyGr2eoF9boZcn/er9I/OVMwD2YyNgAjaYFjWY3rMqdCdcrJoN1TezIX6sqZHcvIggkwV+SD0DgQrS6KlvbP1QuoaCxySiUp737p3axabZftIlLL79MiJOjhcSjXj7PRh2Stkc+TC2XPT7Soj9InkxuYURdWKpy77sfB9qRnULCSCBeP5GRg7Aclqks8hapLghoPg5JL9VjwQmF7rgduQd8HiCq4DcIKU+kPZVJSJluwn9ugWlJlM2XlF6brEIo2qqQDNTtdiWrTP8n4HEVCtIC0BRCcVd22ft50dDkZyqJXUnT6CLevS9ZOrozroKZ1+W8cerjS40SYqnMCdA1bZXfKZMjCkSxMwWJta7SV5Xh4jcJUSZkRnSO0HakNL8QtTa1id1LYeZgq6y9uACAE= ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:53:40.204 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-07T07:53:40.209 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:40 vm07 bash[17031]: audit 2026-03-07T06:53:39.032698+0000 mgr.vm07.yrfcuj (mgr.14168) 9 : audit [DBG] from='client.14182 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:40.209 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:40 vm07 bash[17031]: audit 2026-03-07T06:53:39.032698+0000 mgr.vm07.yrfcuj (mgr.14168) 9 : audit [DBG] from='client.14182 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:40.209 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:40 vm07 bash[17031]: audit 2026-03-07T06:53:39.524381+0000 mon.vm07 (mon.0) 99 : audit [DBG] from='client.? 192.168.123.107:0/443094592' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-07T07:53:40.209 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:40 vm07 bash[17031]: audit 2026-03-07T06:53:39.524381+0000 mon.vm07 (mon.0) 99 : audit [DBG] from='client.? 192.168.123.107:0/443094592' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-07T07:53:40.209 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:40 vm07 bash[17031]: audit 2026-03-07T06:53:39.924928+0000 mon.vm07 (mon.0) 100 : audit [INF] from='client.? 192.168.123.107:0/1117924743' entity='client.admin' 2026-03-07T07:53:40.209 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:40 vm07 bash[17031]: audit 2026-03-07T06:53:39.924928+0000 mon.vm07 (mon.0) 100 : audit [INF] from='client.? 192.168.123.107:0/1117924743' entity='client.admin' 2026-03-07T07:53:40.209 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:40 vm07 bash[17031]: cluster 2026-03-07T06:53:40.185831+0000 mon.vm07 (mon.0) 101 : cluster [DBG] mgrmap e12: vm07.yrfcuj(active, since 2s) 2026-03-07T07:53:40.209 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:40 vm07 bash[17031]: cluster 2026-03-07T06:53:40.185831+0000 mon.vm07 (mon.0) 101 : cluster [DBG] mgrmap e12: vm07.yrfcuj(active, since 2s) 2026-03-07T07:53:43.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:43 vm07 bash[17031]: audit 2026-03-07T06:53:42.327114+0000 mon.vm07 (mon.0) 102 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:43.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:43 vm07 bash[17031]: audit 2026-03-07T06:53:42.327114+0000 mon.vm07 (mon.0) 102 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:43.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:43 vm07 bash[17031]: audit 2026-03-07T06:53:42.869713+0000 mon.vm07 (mon.0) 103 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:43.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:43 vm07 bash[17031]: audit 2026-03-07T06:53:42.869713+0000 mon.vm07 (mon.0) 103 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:44.191 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:53:44.584 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-07T07:53:44.585 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-07T07:53:44.879 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:44 vm07 bash[17031]: cluster 2026-03-07T06:53:43.873335+0000 mon.vm07 (mon.0) 104 : cluster [DBG] mgrmap e13: vm07.yrfcuj(active, since 6s) 2026-03-07T07:53:44.879 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:44 vm07 bash[17031]: cluster 2026-03-07T06:53:43.873335+0000 mon.vm07 (mon.0) 104 : cluster [DBG] mgrmap e13: vm07.yrfcuj(active, since 6s) 2026-03-07T07:53:44.879 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:44 vm07 bash[17031]: audit 2026-03-07T06:53:44.520072+0000 mon.vm07 (mon.0) 105 : audit [INF] from='client.? 192.168.123.107:0/1299184646' entity='client.admin' 2026-03-07T07:53:44.879 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:44 vm07 bash[17031]: audit 2026-03-07T06:53:44.520072+0000 mon.vm07 (mon.0) 105 : audit [INF] from='client.? 192.168.123.107:0/1299184646' entity='client.admin' 2026-03-07T07:53:49.207 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:53:49.856 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:53:49.857 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:53:49.857 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm09 2026-03-07T07:53:49.857 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-07T07:53:49.857 DEBUG:teuthology.orchestra.run.vm09:> dd of=/etc/ceph/ceph.conf 2026-03-07T07:53:49.860 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-07T07:53:49.860 DEBUG:teuthology.orchestra.run.vm09:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-07T07:53:49.906 INFO:tasks.cephadm:Adding host vm09 to orchestrator... 2026-03-07T07:53:49.906 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph orch host add vm09 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:48.974849+0000 mon.vm07 (mon.0) 106 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:48.974849+0000 mon.vm07 (mon.0) 106 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:48.977381+0000 mon.vm07 (mon.0) 107 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:48.977381+0000 mon.vm07 (mon.0) 107 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:48.978059+0000 mon.vm07 (mon.0) 108 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"}]: dispatch 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:48.978059+0000 mon.vm07 (mon.0) 108 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"}]: dispatch 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:48.980403+0000 mon.vm07 (mon.0) 109 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:48.980403+0000 mon.vm07 (mon.0) 109 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:48.981255+0000 mon.vm07 (mon.0) 110 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm07", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:48.981255+0000 mon.vm07 (mon.0) 110 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm07", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:48.982146+0000 mon.vm07 (mon.0) 111 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm07", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:48.982146+0000 mon.vm07 (mon.0) 111 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm07", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:48.983473+0000 mon.vm07 (mon.0) 112 : audit [DBG] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:48.983473+0000 mon.vm07 (mon.0) 112 : audit [DBG] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: cephadm 2026-03-07T06:53:48.983915+0000 mgr.vm07.yrfcuj (mgr.14168) 10 : cephadm [INF] Deploying daemon ceph-exporter.vm07 on vm07 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: cephadm 2026-03-07T06:53:48.983915+0000 mgr.vm07.yrfcuj (mgr.14168) 10 : cephadm [INF] Deploying daemon ceph-exporter.vm07 on vm07 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:49.626613+0000 mgr.vm07.yrfcuj (mgr.14168) 11 : audit [DBG] from='client.14190 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:49.626613+0000 mgr.vm07.yrfcuj (mgr.14168) 11 : audit [DBG] from='client.14190 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:49.632219+0000 mon.vm07 (mon.0) 113 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:49.632219+0000 mon.vm07 (mon.0) 113 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:50.121 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:49.810654+0000 mon.vm07 (mon.0) 114 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:50.122 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:49.810654+0000 mon.vm07 (mon.0) 114 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:50.122 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:49.814341+0000 mon.vm07 (mon.0) 115 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:50.122 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:49.814341+0000 mon.vm07 (mon.0) 115 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:50.122 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:49.817652+0000 mon.vm07 (mon.0) 116 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:50.122 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:49.817652+0000 mon.vm07 (mon.0) 116 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:50.122 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:49.823978+0000 mon.vm07 (mon.0) 117 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:50.122 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:49.823978+0000 mon.vm07 (mon.0) 117 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:50.122 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:49.826223+0000 mon.vm07 (mon.0) 118 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm07", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-07T07:53:50.122 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:49.826223+0000 mon.vm07 (mon.0) 118 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm07", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-07T07:53:50.122 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:49.827845+0000 mon.vm07 (mon.0) 119 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm07", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-07T07:53:50.122 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:49.827845+0000 mon.vm07 (mon.0) 119 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm07", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-07T07:53:50.122 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:49.839319+0000 mon.vm07 (mon.0) 120 : audit [DBG] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:53:50.122 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:49 vm07 bash[17031]: audit 2026-03-07T06:53:49.839319+0000 mon.vm07 (mon.0) 120 : audit [DBG] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:53:50.562 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:50 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:53:50.875 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:50 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:53:51.129 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:53:51.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: cephadm 2026-03-07T06:53:49.845683+0000 mgr.vm07.yrfcuj (mgr.14168) 12 : cephadm [INF] Deploying daemon crash.vm07 on vm07 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: cephadm 2026-03-07T06:53:49.845683+0000 mgr.vm07.yrfcuj (mgr.14168) 12 : cephadm [INF] Deploying daemon crash.vm07 on vm07 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: audit 2026-03-07T06:53:50.639831+0000 mon.vm07 (mon.0) 121 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: audit 2026-03-07T06:53:50.639831+0000 mon.vm07 (mon.0) 121 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: audit 2026-03-07T06:53:50.642314+0000 mon.vm07 (mon.0) 122 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: audit 2026-03-07T06:53:50.642314+0000 mon.vm07 (mon.0) 122 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: audit 2026-03-07T06:53:50.644731+0000 mon.vm07 (mon.0) 123 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: audit 2026-03-07T06:53:50.644731+0000 mon.vm07 (mon.0) 123 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: audit 2026-03-07T06:53:50.647084+0000 mon.vm07 (mon.0) 124 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: audit 2026-03-07T06:53:50.647084+0000 mon.vm07 (mon.0) 124 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: cephadm 2026-03-07T06:53:50.647983+0000 mgr.vm07.yrfcuj (mgr.14168) 13 : cephadm [INF] Deploying daemon node-exporter.vm07 on vm07 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: cephadm 2026-03-07T06:53:50.647983+0000 mgr.vm07.yrfcuj (mgr.14168) 13 : cephadm [INF] Deploying daemon node-exporter.vm07 on vm07 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: audit 2026-03-07T06:53:51.314775+0000 mon.vm07 (mon.0) 125 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: audit 2026-03-07T06:53:51.314775+0000 mon.vm07 (mon.0) 125 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: audit 2026-03-07T06:53:51.317331+0000 mon.vm07 (mon.0) 126 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: audit 2026-03-07T06:53:51.317331+0000 mon.vm07 (mon.0) 126 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: audit 2026-03-07T06:53:51.319555+0000 mon.vm07 (mon.0) 127 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: audit 2026-03-07T06:53:51.319555+0000 mon.vm07 (mon.0) 127 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: audit 2026-03-07T06:53:51.321524+0000 mon.vm07 (mon.0) 128 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:51.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:51 vm07 bash[17031]: audit 2026-03-07T06:53:51.321524+0000 mon.vm07 (mon.0) 128 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:52 vm07 bash[17031]: cephadm 2026-03-07T06:53:51.327392+0000 mgr.vm07.yrfcuj (mgr.14168) 14 : cephadm [INF] Deploying daemon alertmanager.vm07 on vm07 2026-03-07T07:53:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:52 vm07 bash[17031]: cephadm 2026-03-07T06:53:51.327392+0000 mgr.vm07.yrfcuj (mgr.14168) 14 : cephadm [INF] Deploying daemon alertmanager.vm07 on vm07 2026-03-07T07:53:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:52 vm07 bash[17031]: audit 2026-03-07T06:53:52.228458+0000 mon.vm07 (mon.0) 129 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:52 vm07 bash[17031]: audit 2026-03-07T06:53:52.228458+0000 mon.vm07 (mon.0) 129 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:55.710 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:53:55.947 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:55 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:53:55.947 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:55 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:53:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:55.930750+0000 mon.vm07 (mon.0) 130 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:55.930750+0000 mon.vm07 (mon.0) 130 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:55.933781+0000 mon.vm07 (mon.0) 131 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:55.933781+0000 mon.vm07 (mon.0) 131 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:55.936949+0000 mon.vm07 (mon.0) 132 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:55.936949+0000 mon.vm07 (mon.0) 132 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:55.939129+0000 mon.vm07 (mon.0) 133 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:55.939129+0000 mon.vm07 (mon.0) 133 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:55.941878+0000 mon.vm07 (mon.0) 134 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:55.941878+0000 mon.vm07 (mon.0) 134 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:55.948367+0000 mon.vm07 (mon.0) 135 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:55.948367+0000 mon.vm07 (mon.0) 135 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: cephadm 2026-03-07T06:53:55.954287+0000 mgr.vm07.yrfcuj (mgr.14168) 15 : cephadm [INF] Regenerating cephadm self-signed grafana TLS certificates 2026-03-07T07:53:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: cephadm 2026-03-07T06:53:55.954287+0000 mgr.vm07.yrfcuj (mgr.14168) 15 : cephadm [INF] Regenerating cephadm self-signed grafana TLS certificates 2026-03-07T07:53:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:55.995008+0000 mon.vm07 (mon.0) 136 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:55.995008+0000 mon.vm07 (mon.0) 136 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:56.000450+0000 mon.vm07 (mon.0) 137 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:56.000450+0000 mon.vm07 (mon.0) 137 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:56.008467+0000 mon.vm07 (mon.0) 138 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-07T07:53:57.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:56.008467+0000 mon.vm07 (mon.0) 138 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-07T07:53:57.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:56.008730+0000 mgr.vm07.yrfcuj (mgr.14168) 16 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-07T07:53:57.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:56.008730+0000 mgr.vm07.yrfcuj (mgr.14168) 16 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-07T07:53:57.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:56.012740+0000 mon.vm07 (mon.0) 139 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:57.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:56.012740+0000 mon.vm07 (mon.0) 139 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:57.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: cephadm 2026-03-07T06:53:56.020242+0000 mgr.vm07.yrfcuj (mgr.14168) 17 : cephadm [INF] Deploying daemon grafana.vm07 on vm07 2026-03-07T07:53:57.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: cephadm 2026-03-07T06:53:56.020242+0000 mgr.vm07.yrfcuj (mgr.14168) 17 : cephadm [INF] Deploying daemon grafana.vm07 on vm07 2026-03-07T07:53:57.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:56.456902+0000 mgr.vm07.yrfcuj (mgr.14168) 18 : audit [DBG] from='client.14193 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:57.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:56 vm07 bash[17031]: audit 2026-03-07T06:53:56.456902+0000 mgr.vm07.yrfcuj (mgr.14168) 18 : audit [DBG] from='client.14193 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:53:58.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:58 vm07 bash[17031]: cluster 2026-03-07T06:53:57.202491+0000 mgr.vm07.yrfcuj (mgr.14168) 19 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:53:58.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:58 vm07 bash[17031]: cluster 2026-03-07T06:53:57.202491+0000 mgr.vm07.yrfcuj (mgr.14168) 19 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:53:58.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:58 vm07 bash[17031]: audit 2026-03-07T06:53:57.231805+0000 mon.vm07 (mon.0) 140 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:58.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:58 vm07 bash[17031]: audit 2026-03-07T06:53:57.231805+0000 mon.vm07 (mon.0) 140 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:53:58.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:58 vm07 bash[17031]: cephadm 2026-03-07T06:53:57.287542+0000 mgr.vm07.yrfcuj (mgr.14168) 20 : cephadm [INF] Deploying cephadm binary to vm09 2026-03-07T07:53:58.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:58 vm07 bash[17031]: cephadm 2026-03-07T06:53:57.287542+0000 mgr.vm07.yrfcuj (mgr.14168) 20 : cephadm [INF] Deploying cephadm binary to vm09 2026-03-07T07:53:59.002 INFO:teuthology.orchestra.run.vm07.stdout:Added host 'vm09' with addr '192.168.123.109' 2026-03-07T07:53:59.062 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph orch host ls --format=json 2026-03-07T07:54:00.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:59 vm07 bash[17031]: audit 2026-03-07T06:53:58.993904+0000 mon.vm07 (mon.0) 141 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:00.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:59 vm07 bash[17031]: audit 2026-03-07T06:53:58.993904+0000 mon.vm07 (mon.0) 141 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:00.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:59 vm07 bash[17031]: cephadm 2026-03-07T06:53:58.994261+0000 mgr.vm07.yrfcuj (mgr.14168) 21 : cephadm [INF] Added host vm09 2026-03-07T07:54:00.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:59 vm07 bash[17031]: cephadm 2026-03-07T06:53:58.994261+0000 mgr.vm07.yrfcuj (mgr.14168) 21 : cephadm [INF] Added host vm09 2026-03-07T07:54:00.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:59 vm07 bash[17031]: cluster 2026-03-07T06:53:59.202668+0000 mgr.vm07.yrfcuj (mgr.14168) 22 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:00.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:53:59 vm07 bash[17031]: cluster 2026-03-07T06:53:59.202668+0000 mgr.vm07.yrfcuj (mgr.14168) 22 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:02.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:02 vm07 bash[17031]: cluster 2026-03-07T06:54:01.202849+0000 mgr.vm07.yrfcuj (mgr.14168) 23 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:02.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:02 vm07 bash[17031]: cluster 2026-03-07T06:54:01.202849+0000 mgr.vm07.yrfcuj (mgr.14168) 23 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:03.851 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:54:04.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:04 vm07 bash[17031]: cluster 2026-03-07T06:54:03.203075+0000 mgr.vm07.yrfcuj (mgr.14168) 24 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:04.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:04 vm07 bash[17031]: cluster 2026-03-07T06:54:03.203075+0000 mgr.vm07.yrfcuj (mgr.14168) 24 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:04.741 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:54:04.742 INFO:teuthology.orchestra.run.vm07.stdout:[{"addr": "192.168.123.107", "hostname": "vm07", "labels": [], "status": ""}, {"addr": "192.168.123.109", "hostname": "vm09", "labels": [], "status": ""}] 2026-03-07T07:54:04.814 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-07T07:54:04.814 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd crush tunables default 2026-03-07T07:54:05.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:05 vm07 bash[17031]: audit 2026-03-07T06:54:04.736180+0000 mgr.vm07.yrfcuj (mgr.14168) 25 : audit [DBG] from='client.14195 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:54:05.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:05 vm07 bash[17031]: audit 2026-03-07T06:54:04.736180+0000 mgr.vm07.yrfcuj (mgr.14168) 25 : audit [DBG] from='client.14195 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:54:06.620 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:06 vm07 bash[17031]: cluster 2026-03-07T06:54:05.203268+0000 mgr.vm07.yrfcuj (mgr.14168) 26 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:06.620 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:06 vm07 bash[17031]: cluster 2026-03-07T06:54:05.203268+0000 mgr.vm07.yrfcuj (mgr.14168) 26 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:07.735 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:07 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:54:07.735 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:07 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:54:08.738 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:08 vm07 bash[17031]: cluster 2026-03-07T06:54:07.203430+0000 mgr.vm07.yrfcuj (mgr.14168) 27 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:08.738 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:08 vm07 bash[17031]: cluster 2026-03-07T06:54:07.203430+0000 mgr.vm07.yrfcuj (mgr.14168) 27 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:08.738 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:08 vm07 bash[17031]: audit 2026-03-07T06:54:07.711749+0000 mon.vm07 (mon.0) 142 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:08.738 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:08 vm07 bash[17031]: audit 2026-03-07T06:54:07.711749+0000 mon.vm07 (mon.0) 142 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:08.738 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:08 vm07 bash[17031]: audit 2026-03-07T06:54:07.714559+0000 mon.vm07 (mon.0) 143 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:08.738 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:08 vm07 bash[17031]: audit 2026-03-07T06:54:07.714559+0000 mon.vm07 (mon.0) 143 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:08.739 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:08 vm07 bash[17031]: audit 2026-03-07T06:54:07.716819+0000 mon.vm07 (mon.0) 144 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:08.739 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:08 vm07 bash[17031]: audit 2026-03-07T06:54:07.716819+0000 mon.vm07 (mon.0) 144 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:08.739 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:08 vm07 bash[17031]: audit 2026-03-07T06:54:07.718541+0000 mon.vm07 (mon.0) 145 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:08.739 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:08 vm07 bash[17031]: audit 2026-03-07T06:54:07.718541+0000 mon.vm07 (mon.0) 145 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:08.739 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:08 vm07 bash[17031]: audit 2026-03-07T06:54:07.720341+0000 mon.vm07 (mon.0) 146 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:08.739 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:08 vm07 bash[17031]: audit 2026-03-07T06:54:07.720341+0000 mon.vm07 (mon.0) 146 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:08.739 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:08 vm07 bash[17031]: audit 2026-03-07T06:54:07.722746+0000 mon.vm07 (mon.0) 147 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:08.739 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:08 vm07 bash[17031]: audit 2026-03-07T06:54:07.722746+0000 mon.vm07 (mon.0) 147 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:08.739 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:08 vm07 bash[17031]: audit 2026-03-07T06:54:07.724649+0000 mon.vm07 (mon.0) 148 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:08.739 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:08 vm07 bash[17031]: audit 2026-03-07T06:54:07.724649+0000 mon.vm07 (mon.0) 148 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:08.739 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:08 vm07 bash[17031]: audit 2026-03-07T06:54:07.727680+0000 mon.vm07 (mon.0) 149 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:08.739 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:08 vm07 bash[17031]: audit 2026-03-07T06:54:07.727680+0000 mon.vm07 (mon.0) 149 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:09.568 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:54:09.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:09 vm07 bash[17031]: cephadm 2026-03-07T06:54:07.899791+0000 mgr.vm07.yrfcuj (mgr.14168) 28 : cephadm [INF] Deploying daemon prometheus.vm07 on vm07 2026-03-07T07:54:09.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:09 vm07 bash[17031]: cephadm 2026-03-07T06:54:07.899791+0000 mgr.vm07.yrfcuj (mgr.14168) 28 : cephadm [INF] Deploying daemon prometheus.vm07 on vm07 2026-03-07T07:54:10.728 INFO:teuthology.orchestra.run.vm07.stderr:adjusted tunables profile to default 2026-03-07T07:54:10.832 INFO:tasks.cephadm:Adding mon.vm07 on vm07 2026-03-07T07:54:10.832 INFO:tasks.cephadm:Adding mon.vm09 on vm09 2026-03-07T07:54:10.833 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph orch apply mon '2;vm07:192.168.123.107=vm07;vm09:192.168.123.109=vm09' 2026-03-07T07:54:11.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:10 vm07 bash[17031]: cluster 2026-03-07T06:54:09.203653+0000 mgr.vm07.yrfcuj (mgr.14168) 29 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:11.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:10 vm07 bash[17031]: cluster 2026-03-07T06:54:09.203653+0000 mgr.vm07.yrfcuj (mgr.14168) 29 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:11.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:10 vm07 bash[17031]: audit 2026-03-07T06:54:09.962669+0000 mon.vm07 (mon.0) 150 : audit [INF] from='client.? 192.168.123.107:0/1489886866' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-07T07:54:11.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:10 vm07 bash[17031]: audit 2026-03-07T06:54:09.962669+0000 mon.vm07 (mon.0) 150 : audit [INF] from='client.? 192.168.123.107:0/1489886866' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-07T07:54:12.084 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-07T07:54:12.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:11 vm07 bash[17031]: audit 2026-03-07T06:54:10.720084+0000 mon.vm07 (mon.0) 151 : audit [INF] from='client.? 192.168.123.107:0/1489886866' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-07T07:54:12.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:11 vm07 bash[17031]: audit 2026-03-07T06:54:10.720084+0000 mon.vm07 (mon.0) 151 : audit [INF] from='client.? 192.168.123.107:0/1489886866' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-07T07:54:12.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:11 vm07 bash[17031]: cluster 2026-03-07T06:54:10.721751+0000 mon.vm07 (mon.0) 152 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-07T07:54:12.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:11 vm07 bash[17031]: cluster 2026-03-07T06:54:10.721751+0000 mon.vm07 (mon.0) 152 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-07T07:54:13.104 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-07T07:54:13.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:12 vm07 bash[17031]: cluster 2026-03-07T06:54:11.203954+0000 mgr.vm07.yrfcuj (mgr.14168) 30 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:13.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:12 vm07 bash[17031]: cluster 2026-03-07T06:54:11.203954+0000 mgr.vm07.yrfcuj (mgr.14168) 30 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:13.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:12 vm07 bash[17031]: audit 2026-03-07T06:54:12.237033+0000 mon.vm07 (mon.0) 153 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:13.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:12 vm07 bash[17031]: audit 2026-03-07T06:54:12.237033+0000 mon.vm07 (mon.0) 153 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:13.799 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled mon update... 2026-03-07T07:54:13.867 DEBUG:teuthology.orchestra.run.vm09:mon.vm09> sudo journalctl -f -n 0 -u ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@mon.vm09.service 2026-03-07T07:54:13.868 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-07T07:54:13.868 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph mon dump -f json 2026-03-07T07:54:14.392 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:14 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:54:14.392 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:14 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:54:15.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:14 vm07 bash[17031]: cluster 2026-03-07T06:54:13.204213+0000 mgr.vm07.yrfcuj (mgr.14168) 31 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:15.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:14 vm07 bash[17031]: cluster 2026-03-07T06:54:13.204213+0000 mgr.vm07.yrfcuj (mgr.14168) 31 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:15.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:14 vm07 bash[17031]: audit 2026-03-07T06:54:13.492338+0000 mgr.vm07.yrfcuj (mgr.14168) 32 : audit [DBG] from='client.14199 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm07:192.168.123.107=vm07;vm09:192.168.123.109=vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:54:15.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:14 vm07 bash[17031]: audit 2026-03-07T06:54:13.492338+0000 mgr.vm07.yrfcuj (mgr.14168) 32 : audit [DBG] from='client.14199 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm07:192.168.123.107=vm07;vm09:192.168.123.109=vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:54:15.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:14 vm07 bash[17031]: cephadm 2026-03-07T06:54:13.493461+0000 mgr.vm07.yrfcuj (mgr.14168) 33 : cephadm [INF] Saving service mon spec with placement vm07:192.168.123.107=vm07;vm09:192.168.123.109=vm09;count:2 2026-03-07T07:54:15.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:14 vm07 bash[17031]: cephadm 2026-03-07T06:54:13.493461+0000 mgr.vm07.yrfcuj (mgr.14168) 33 : cephadm [INF] Saving service mon spec with placement vm07:192.168.123.107=vm07;vm09:192.168.123.109=vm09;count:2 2026-03-07T07:54:15.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:14 vm07 bash[17031]: audit 2026-03-07T06:54:13.790730+0000 mon.vm07 (mon.0) 154 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:15.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:14 vm07 bash[17031]: audit 2026-03-07T06:54:13.790730+0000 mon.vm07 (mon.0) 154 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:15.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:14 vm07 bash[17031]: audit 2026-03-07T06:54:14.421526+0000 mon.vm07 (mon.0) 155 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:15.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:14 vm07 bash[17031]: audit 2026-03-07T06:54:14.421526+0000 mon.vm07 (mon.0) 155 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:15.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:14 vm07 bash[17031]: audit 2026-03-07T06:54:14.424296+0000 mon.vm07 (mon.0) 156 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:15.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:14 vm07 bash[17031]: audit 2026-03-07T06:54:14.424296+0000 mon.vm07 (mon.0) 156 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:15.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:14 vm07 bash[17031]: audit 2026-03-07T06:54:14.426689+0000 mon.vm07 (mon.0) 157 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:15.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:14 vm07 bash[17031]: audit 2026-03-07T06:54:14.426689+0000 mon.vm07 (mon.0) 157 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:15.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:14 vm07 bash[17031]: audit 2026-03-07T06:54:14.427819+0000 mon.vm07 (mon.0) 158 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr module enable", "module": "prometheus"}]: dispatch 2026-03-07T07:54:15.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:14 vm07 bash[17031]: audit 2026-03-07T06:54:14.427819+0000 mon.vm07 (mon.0) 158 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr module enable", "module": "prometheus"}]: dispatch 2026-03-07T07:54:15.171 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-07T07:54:16.192 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-07T07:54:16.584 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-07T07:54:16.584 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-07T07:54:16.584 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"312fdbc4-19f2-11f1-81d0-4bbd10a1e012","modified":"2026-03-07T06:52:50.872627Z","created":"2026-03-07T06:52:50.872627Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm07","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:3300","nonce":0},{"type":"v1","addr":"192.168.123.107:6789","nonce":0}]},"addr":"192.168.123.107:6789/0","public_addr":"192.168.123.107:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-07T07:54:16.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:16 vm07 bash[17031]: audit 2026-03-07T06:54:15.428919+0000 mon.vm07 (mon.0) 159 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "mgr module enable", "module": "prometheus"}]': finished 2026-03-07T07:54:16.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:16 vm07 bash[17031]: audit 2026-03-07T06:54:15.428919+0000 mon.vm07 (mon.0) 159 : audit [INF] from='mgr.14168 192.168.123.107:0/418138352' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "mgr module enable", "module": "prometheus"}]': finished 2026-03-07T07:54:16.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:16 vm07 bash[17031]: cluster 2026-03-07T06:54:15.432359+0000 mon.vm07 (mon.0) 160 : cluster [DBG] mgrmap e14: vm07.yrfcuj(active, since 38s) 2026-03-07T07:54:16.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:16 vm07 bash[17031]: cluster 2026-03-07T06:54:15.432359+0000 mon.vm07 (mon.0) 160 : cluster [DBG] mgrmap e14: vm07.yrfcuj(active, since 38s) 2026-03-07T07:54:17.653 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-07T07:54:17.653 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph mon dump -f json 2026-03-07T07:54:17.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:17 vm07 bash[17031]: audit 2026-03-07T06:54:16.578771+0000 mon.vm07 (mon.0) 161 : audit [DBG] from='client.? 192.168.123.109:0/718001142' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T07:54:17.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:17 vm07 bash[17031]: audit 2026-03-07T06:54:16.578771+0000 mon.vm07 (mon.0) 161 : audit [DBG] from='client.? 192.168.123.109:0/718001142' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T07:54:18.898 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-07T07:54:19.919 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-07T07:54:20.305 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-07T07:54:20.306 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"312fdbc4-19f2-11f1-81d0-4bbd10a1e012","modified":"2026-03-07T06:52:50.872627Z","created":"2026-03-07T06:52:50.872627Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm07","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:3300","nonce":0},{"type":"v1","addr":"192.168.123.107:6789","nonce":0}]},"addr":"192.168.123.107:6789/0","public_addr":"192.168.123.107:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-07T07:54:20.306 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-07T07:54:20.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:20 vm07 bash[17031]: audit 2026-03-07T06:54:20.300050+0000 mon.vm07 (mon.0) 162 : audit [DBG] from='client.? 192.168.123.109:0/3560125010' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T07:54:20.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:20 vm07 bash[17031]: audit 2026-03-07T06:54:20.300050+0000 mon.vm07 (mon.0) 162 : audit [DBG] from='client.? 192.168.123.109:0/3560125010' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T07:54:21.384 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-07T07:54:21.384 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph mon dump -f json 2026-03-07T07:54:22.630 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-07T07:54:23.651 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-07T07:54:24.012 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-07T07:54:24.012 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"312fdbc4-19f2-11f1-81d0-4bbd10a1e012","modified":"2026-03-07T06:52:50.872627Z","created":"2026-03-07T06:52:50.872627Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm07","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:3300","nonce":0},{"type":"v1","addr":"192.168.123.107:6789","nonce":0}]},"addr":"192.168.123.107:6789/0","public_addr":"192.168.123.107:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-07T07:54:24.012 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-07T07:54:24.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:24 vm07 bash[17031]: audit 2026-03-07T06:54:24.006928+0000 mon.vm07 (mon.0) 163 : audit [DBG] from='client.? 192.168.123.109:0/1758613084' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T07:54:24.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:24 vm07 bash[17031]: audit 2026-03-07T06:54:24.006928+0000 mon.vm07 (mon.0) 163 : audit [DBG] from='client.? 192.168.123.109:0/1758613084' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T07:54:25.074 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-07T07:54:25.074 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph mon dump -f json 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: cluster 2026-03-07T06:54:25.066428+0000 mon.vm07 (mon.0) 164 : cluster [INF] Active manager daemon vm07.yrfcuj restarted 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: cluster 2026-03-07T06:54:25.066428+0000 mon.vm07 (mon.0) 164 : cluster [INF] Active manager daemon vm07.yrfcuj restarted 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: cluster 2026-03-07T06:54:25.066815+0000 mon.vm07 (mon.0) 165 : cluster [INF] Activating manager daemon vm07.yrfcuj 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: cluster 2026-03-07T06:54:25.066815+0000 mon.vm07 (mon.0) 165 : cluster [INF] Activating manager daemon vm07.yrfcuj 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: cluster 2026-03-07T06:54:25.071765+0000 mon.vm07 (mon.0) 166 : cluster [DBG] osdmap e5: 0 total, 0 up, 0 in 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: cluster 2026-03-07T06:54:25.071765+0000 mon.vm07 (mon.0) 166 : cluster [DBG] osdmap e5: 0 total, 0 up, 0 in 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: cluster 2026-03-07T06:54:25.071905+0000 mon.vm07 (mon.0) 167 : cluster [DBG] mgrmap e15: vm07.yrfcuj(active, starting, since 0.00517675s) 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: cluster 2026-03-07T06:54:25.071905+0000 mon.vm07 (mon.0) 167 : cluster [DBG] mgrmap e15: vm07.yrfcuj(active, starting, since 0.00517675s) 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: audit 2026-03-07T06:54:25.075249+0000 mon.vm07 (mon.0) 168 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: audit 2026-03-07T06:54:25.075249+0000 mon.vm07 (mon.0) 168 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: audit 2026-03-07T06:54:25.075666+0000 mon.vm07 (mon.0) 169 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr metadata", "who": "vm07.yrfcuj", "id": "vm07.yrfcuj"}]: dispatch 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: audit 2026-03-07T06:54:25.075666+0000 mon.vm07 (mon.0) 169 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr metadata", "who": "vm07.yrfcuj", "id": "vm07.yrfcuj"}]: dispatch 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: audit 2026-03-07T06:54:25.076891+0000 mon.vm07 (mon.0) 170 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: audit 2026-03-07T06:54:25.076891+0000 mon.vm07 (mon.0) 170 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: audit 2026-03-07T06:54:25.077509+0000 mon.vm07 (mon.0) 171 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: audit 2026-03-07T06:54:25.077509+0000 mon.vm07 (mon.0) 171 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: audit 2026-03-07T06:54:25.077879+0000 mon.vm07 (mon.0) 172 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: audit 2026-03-07T06:54:25.077879+0000 mon.vm07 (mon.0) 172 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: cluster 2026-03-07T06:54:25.084673+0000 mon.vm07 (mon.0) 173 : cluster [INF] Manager daemon vm07.yrfcuj is now available 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: cluster 2026-03-07T06:54:25.084673+0000 mon.vm07 (mon.0) 173 : cluster [INF] Manager daemon vm07.yrfcuj is now available 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: audit 2026-03-07T06:54:25.102418+0000 mon.vm07 (mon.0) 174 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: audit 2026-03-07T06:54:25.102418+0000 mon.vm07 (mon.0) 174 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: audit 2026-03-07T06:54:25.115924+0000 mon.vm07 (mon.0) 175 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:54:25.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:25 vm07 bash[17031]: audit 2026-03-07T06:54:25.115924+0000 mon.vm07 (mon.0) 175 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:54:26.343 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-07T07:54:26.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:26 vm07 bash[17031]: audit 2026-03-07T06:54:25.130902+0000 mon.vm07 (mon.0) 176 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:54:26.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:26 vm07 bash[17031]: audit 2026-03-07T06:54:25.130902+0000 mon.vm07 (mon.0) 176 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:54:26.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:26 vm07 bash[17031]: audit 2026-03-07T06:54:25.131864+0000 mon.vm07 (mon.0) 177 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/mirror_snapshot_schedule"}]: dispatch 2026-03-07T07:54:26.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:26 vm07 bash[17031]: audit 2026-03-07T06:54:25.131864+0000 mon.vm07 (mon.0) 177 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/mirror_snapshot_schedule"}]: dispatch 2026-03-07T07:54:26.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:26 vm07 bash[17031]: audit 2026-03-07T06:54:25.167407+0000 mon.vm07 (mon.0) 178 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/trash_purge_schedule"}]: dispatch 2026-03-07T07:54:26.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:26 vm07 bash[17031]: audit 2026-03-07T06:54:25.167407+0000 mon.vm07 (mon.0) 178 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/trash_purge_schedule"}]: dispatch 2026-03-07T07:54:26.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:26 vm07 bash[17031]: audit 2026-03-07T06:54:25.739320+0000 mon.vm07 (mon.0) 179 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:26.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:26 vm07 bash[17031]: audit 2026-03-07T06:54:25.739320+0000 mon.vm07 (mon.0) 179 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:26.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:26 vm07 bash[17031]: cluster 2026-03-07T06:54:26.075119+0000 mon.vm07 (mon.0) 180 : cluster [DBG] mgrmap e16: vm07.yrfcuj(active, since 1.00839s) 2026-03-07T07:54:26.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:26 vm07 bash[17031]: cluster 2026-03-07T06:54:26.075119+0000 mon.vm07 (mon.0) 180 : cluster [DBG] mgrmap e16: vm07.yrfcuj(active, since 1.00839s) 2026-03-07T07:54:27.361 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-07T07:54:27.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:27 vm07 bash[17031]: cephadm 2026-03-07T06:54:25.998985+0000 mgr.vm07.yrfcuj (mgr.14201) 1 : cephadm [INF] [07/Mar/2026:06:54:25] ENGINE Bus STARTING 2026-03-07T07:54:27.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:27 vm07 bash[17031]: cephadm 2026-03-07T06:54:25.998985+0000 mgr.vm07.yrfcuj (mgr.14201) 1 : cephadm [INF] [07/Mar/2026:06:54:25] ENGINE Bus STARTING 2026-03-07T07:54:27.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:27 vm07 bash[17031]: cephadm 2026-03-07T06:54:26.101321+0000 mgr.vm07.yrfcuj (mgr.14201) 2 : cephadm [INF] [07/Mar/2026:06:54:26] ENGINE Serving on http://192.168.123.107:8765 2026-03-07T07:54:27.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:27 vm07 bash[17031]: cephadm 2026-03-07T06:54:26.101321+0000 mgr.vm07.yrfcuj (mgr.14201) 2 : cephadm [INF] [07/Mar/2026:06:54:26] ENGINE Serving on http://192.168.123.107:8765 2026-03-07T07:54:27.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:27 vm07 bash[17031]: cephadm 2026-03-07T06:54:26.210153+0000 mgr.vm07.yrfcuj (mgr.14201) 3 : cephadm [INF] [07/Mar/2026:06:54:26] ENGINE Serving on https://192.168.123.107:7150 2026-03-07T07:54:27.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:27 vm07 bash[17031]: cephadm 2026-03-07T06:54:26.210153+0000 mgr.vm07.yrfcuj (mgr.14201) 3 : cephadm [INF] [07/Mar/2026:06:54:26] ENGINE Serving on https://192.168.123.107:7150 2026-03-07T07:54:27.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:27 vm07 bash[17031]: cephadm 2026-03-07T06:54:26.210271+0000 mgr.vm07.yrfcuj (mgr.14201) 4 : cephadm [INF] [07/Mar/2026:06:54:26] ENGINE Bus STARTED 2026-03-07T07:54:27.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:27 vm07 bash[17031]: cephadm 2026-03-07T06:54:26.210271+0000 mgr.vm07.yrfcuj (mgr.14201) 4 : cephadm [INF] [07/Mar/2026:06:54:26] ENGINE Bus STARTED 2026-03-07T07:54:27.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:27 vm07 bash[17031]: cephadm 2026-03-07T06:54:26.210489+0000 mgr.vm07.yrfcuj (mgr.14201) 5 : cephadm [INF] [07/Mar/2026:06:54:26] ENGINE Client ('192.168.123.107', 47182) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-07T07:54:27.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:27 vm07 bash[17031]: cephadm 2026-03-07T06:54:26.210489+0000 mgr.vm07.yrfcuj (mgr.14201) 5 : cephadm [INF] [07/Mar/2026:06:54:26] ENGINE Client ('192.168.123.107', 47182) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-07T07:54:27.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:27 vm07 bash[17031]: audit 2026-03-07T06:54:27.022992+0000 mon.vm07 (mon.0) 181 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:27.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:27 vm07 bash[17031]: audit 2026-03-07T06:54:27.022992+0000 mon.vm07 (mon.0) 181 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:27.751 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-07T07:54:27.751 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"312fdbc4-19f2-11f1-81d0-4bbd10a1e012","modified":"2026-03-07T06:52:50.872627Z","created":"2026-03-07T06:52:50.872627Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm07","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:3300","nonce":0},{"type":"v1","addr":"192.168.123.107:6789","nonce":0}]},"addr":"192.168.123.107:6789/0","public_addr":"192.168.123.107:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-07T07:54:27.751 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-07T07:54:28.822 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-07T07:54:28.822 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph mon dump -f json 2026-03-07T07:54:28.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:28 vm07 bash[17031]: audit 2026-03-07T06:54:27.615589+0000 mon.vm07 (mon.0) 182 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:28.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:28 vm07 bash[17031]: audit 2026-03-07T06:54:27.615589+0000 mon.vm07 (mon.0) 182 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:28.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:28 vm07 bash[17031]: audit 2026-03-07T06:54:27.745479+0000 mon.vm07 (mon.0) 183 : audit [DBG] from='client.? 192.168.123.109:0/1396054845' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T07:54:28.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:28 vm07 bash[17031]: audit 2026-03-07T06:54:27.745479+0000 mon.vm07 (mon.0) 183 : audit [DBG] from='client.? 192.168.123.109:0/1396054845' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T07:54:28.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:28 vm07 bash[17031]: cluster 2026-03-07T06:54:28.027190+0000 mon.vm07 (mon.0) 184 : cluster [DBG] mgrmap e17: vm07.yrfcuj(active, since 2s) 2026-03-07T07:54:28.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:28 vm07 bash[17031]: cluster 2026-03-07T06:54:28.027190+0000 mon.vm07 (mon.0) 184 : cluster [DBG] mgrmap e17: vm07.yrfcuj(active, since 2s) 2026-03-07T07:54:31.150 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-07T07:54:32.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:30.668232+0000 mon.vm07 (mon.0) 185 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:30.668232+0000 mon.vm07 (mon.0) 185 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:30.670781+0000 mon.vm07 (mon.0) 186 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:30.670781+0000 mon.vm07 (mon.0) 186 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:30.673303+0000 mon.vm07 (mon.0) 187 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:30.673303+0000 mon.vm07 (mon.0) 187 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:30.675499+0000 mon.vm07 (mon.0) 188 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:30.675499+0000 mon.vm07 (mon.0) 188 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:30.676021+0000 mon.vm07 (mon.0) 189 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-07T07:54:32.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:30.676021+0000 mon.vm07 (mon.0) 189 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-07T07:54:32.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:30.972282+0000 mon.vm07 (mon.0) 190 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:30.972282+0000 mon.vm07 (mon.0) 190 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:30.974998+0000 mon.vm07 (mon.0) 191 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:30.974998+0000 mon.vm07 (mon.0) 191 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.488452+0000 mon.vm07 (mon.0) 192 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.488452+0000 mon.vm07 (mon.0) 192 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.490892+0000 mon.vm07 (mon.0) 193 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.490892+0000 mon.vm07 (mon.0) 193 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.491488+0000 mon.vm07 (mon.0) 194 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"}]: dispatch 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.491488+0000 mon.vm07 (mon.0) 194 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"}]: dispatch 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.492035+0000 mon.vm07 (mon.0) 195 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.492035+0000 mon.vm07 (mon.0) 195 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.492359+0000 mon.vm07 (mon.0) 196 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.492359+0000 mon.vm07 (mon.0) 196 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.625758+0000 mon.vm07 (mon.0) 197 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.625758+0000 mon.vm07 (mon.0) 197 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.627957+0000 mon.vm07 (mon.0) 198 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.627957+0000 mon.vm07 (mon.0) 198 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.629725+0000 mon.vm07 (mon.0) 199 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.629725+0000 mon.vm07 (mon.0) 199 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.631776+0000 mon.vm07 (mon.0) 200 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.631776+0000 mon.vm07 (mon.0) 200 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.633777+0000 mon.vm07 (mon.0) 201 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.633777+0000 mon.vm07 (mon.0) 201 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.634569+0000 mon.vm07 (mon.0) 202 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.634569+0000 mon.vm07 (mon.0) 202 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.635654+0000 mon.vm07 (mon.0) 203 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.635654+0000 mon.vm07 (mon.0) 203 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.636763+0000 mon.vm07 (mon.0) 204 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:32.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:31 vm07 bash[17031]: audit 2026-03-07T06:54:31.636763+0000 mon.vm07 (mon.0) 204 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:32.180 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.conf 2026-03-07T07:54:32.817 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-07T07:54:32.817 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-07T07:54:32.817 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"312fdbc4-19f2-11f1-81d0-4bbd10a1e012","modified":"2026-03-07T06:52:50.872627Z","created":"2026-03-07T06:52:50.872627Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm07","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:3300","nonce":0},{"type":"v1","addr":"192.168.123.107:6789","nonce":0}]},"addr":"192.168.123.107:6789/0","public_addr":"192.168.123.107:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-07T07:54:33.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:32 vm07 bash[17031]: cephadm 2026-03-07T06:54:31.492906+0000 mgr.vm07.yrfcuj (mgr.14201) 6 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-07T07:54:33.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:32 vm07 bash[17031]: cephadm 2026-03-07T06:54:31.492906+0000 mgr.vm07.yrfcuj (mgr.14201) 6 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-07T07:54:33.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:32 vm07 bash[17031]: cephadm 2026-03-07T06:54:31.492986+0000 mgr.vm07.yrfcuj (mgr.14201) 7 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-07T07:54:33.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:32 vm07 bash[17031]: cephadm 2026-03-07T06:54:31.492986+0000 mgr.vm07.yrfcuj (mgr.14201) 7 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-07T07:54:33.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:32 vm07 bash[17031]: cephadm 2026-03-07T06:54:31.525712+0000 mgr.vm07.yrfcuj (mgr.14201) 8 : cephadm [INF] Updating vm09:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.conf 2026-03-07T07:54:33.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:32 vm07 bash[17031]: cephadm 2026-03-07T06:54:31.525712+0000 mgr.vm07.yrfcuj (mgr.14201) 8 : cephadm [INF] Updating vm09:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.conf 2026-03-07T07:54:33.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:32 vm07 bash[17031]: cephadm 2026-03-07T06:54:31.525781+0000 mgr.vm07.yrfcuj (mgr.14201) 9 : cephadm [INF] Updating vm07:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.conf 2026-03-07T07:54:33.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:32 vm07 bash[17031]: cephadm 2026-03-07T06:54:31.525781+0000 mgr.vm07.yrfcuj (mgr.14201) 9 : cephadm [INF] Updating vm07:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.conf 2026-03-07T07:54:33.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:32 vm07 bash[17031]: cephadm 2026-03-07T06:54:31.557402+0000 mgr.vm07.yrfcuj (mgr.14201) 10 : cephadm [INF] Updating vm07:/etc/ceph/ceph.client.admin.keyring 2026-03-07T07:54:33.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:32 vm07 bash[17031]: cephadm 2026-03-07T06:54:31.557402+0000 mgr.vm07.yrfcuj (mgr.14201) 10 : cephadm [INF] Updating vm07:/etc/ceph/ceph.client.admin.keyring 2026-03-07T07:54:33.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:32 vm07 bash[17031]: cephadm 2026-03-07T06:54:31.557648+0000 mgr.vm07.yrfcuj (mgr.14201) 11 : cephadm [INF] Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-07T07:54:33.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:32 vm07 bash[17031]: cephadm 2026-03-07T06:54:31.557648+0000 mgr.vm07.yrfcuj (mgr.14201) 11 : cephadm [INF] Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-07T07:54:33.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:32 vm07 bash[17031]: cephadm 2026-03-07T06:54:31.589971+0000 mgr.vm07.yrfcuj (mgr.14201) 12 : cephadm [INF] Updating vm07:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.client.admin.keyring 2026-03-07T07:54:33.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:32 vm07 bash[17031]: cephadm 2026-03-07T06:54:31.589971+0000 mgr.vm07.yrfcuj (mgr.14201) 12 : cephadm [INF] Updating vm07:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.client.admin.keyring 2026-03-07T07:54:33.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:32 vm07 bash[17031]: cephadm 2026-03-07T06:54:31.590191+0000 mgr.vm07.yrfcuj (mgr.14201) 13 : cephadm [INF] Updating vm09:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.client.admin.keyring 2026-03-07T07:54:33.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:32 vm07 bash[17031]: cephadm 2026-03-07T06:54:31.590191+0000 mgr.vm07.yrfcuj (mgr.14201) 13 : cephadm [INF] Updating vm09:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.client.admin.keyring 2026-03-07T07:54:33.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:32 vm07 bash[17031]: cephadm 2026-03-07T06:54:31.637330+0000 mgr.vm07.yrfcuj (mgr.14201) 14 : cephadm [INF] Deploying daemon ceph-exporter.vm09 on vm09 2026-03-07T07:54:33.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:32 vm07 bash[17031]: cephadm 2026-03-07T06:54:31.637330+0000 mgr.vm07.yrfcuj (mgr.14201) 14 : cephadm [INF] Deploying daemon ceph-exporter.vm09 on vm09 2026-03-07T07:54:33.958 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-07T07:54:33.959 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph mon dump -f json 2026-03-07T07:54:34.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:33 vm07 bash[17031]: audit 2026-03-07T06:54:32.811912+0000 mon.vm07 (mon.0) 205 : audit [DBG] from='client.? 192.168.123.109:0/1490086917' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T07:54:34.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:33 vm07 bash[17031]: audit 2026-03-07T06:54:32.811912+0000 mon.vm07 (mon.0) 205 : audit [DBG] from='client.? 192.168.123.109:0/1490086917' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T07:54:34.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:33 vm07 bash[17031]: audit 2026-03-07T06:54:32.920774+0000 mon.vm07 (mon.0) 206 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:34.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:33 vm07 bash[17031]: audit 2026-03-07T06:54:32.920774+0000 mon.vm07 (mon.0) 206 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:34.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:33 vm07 bash[17031]: audit 2026-03-07T06:54:32.923304+0000 mon.vm07 (mon.0) 207 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:34.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:33 vm07 bash[17031]: audit 2026-03-07T06:54:32.923304+0000 mon.vm07 (mon.0) 207 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:34.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:33 vm07 bash[17031]: audit 2026-03-07T06:54:32.925626+0000 mon.vm07 (mon.0) 208 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:34.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:33 vm07 bash[17031]: audit 2026-03-07T06:54:32.925626+0000 mon.vm07 (mon.0) 208 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:34.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:33 vm07 bash[17031]: audit 2026-03-07T06:54:32.927598+0000 mon.vm07 (mon.0) 209 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:34.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:33 vm07 bash[17031]: audit 2026-03-07T06:54:32.927598+0000 mon.vm07 (mon.0) 209 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:34.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:33 vm07 bash[17031]: audit 2026-03-07T06:54:32.928329+0000 mon.vm07 (mon.0) 210 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-07T07:54:34.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:33 vm07 bash[17031]: audit 2026-03-07T06:54:32.928329+0000 mon.vm07 (mon.0) 210 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-07T07:54:34.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:33 vm07 bash[17031]: audit 2026-03-07T06:54:32.929293+0000 mon.vm07 (mon.0) 211 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-07T07:54:34.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:33 vm07 bash[17031]: audit 2026-03-07T06:54:32.929293+0000 mon.vm07 (mon.0) 211 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-07T07:54:34.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:33 vm07 bash[17031]: audit 2026-03-07T06:54:32.930583+0000 mon.vm07 (mon.0) 212 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:34.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:33 vm07 bash[17031]: audit 2026-03-07T06:54:32.930583+0000 mon.vm07 (mon.0) 212 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:34.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:33 vm07 bash[17031]: cephadm 2026-03-07T06:54:32.931049+0000 mgr.vm07.yrfcuj (mgr.14201) 15 : cephadm [INF] Deploying daemon crash.vm09 on vm09 2026-03-07T07:54:34.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:33 vm07 bash[17031]: cephadm 2026-03-07T06:54:32.931049+0000 mgr.vm07.yrfcuj (mgr.14201) 15 : cephadm [INF] Deploying daemon crash.vm09 on vm09 2026-03-07T07:54:35.051 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:33.720078+0000 mon.vm07 (mon.0) 213 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:35.051 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:33.720078+0000 mon.vm07 (mon.0) 213 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:35.051 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:33.723072+0000 mon.vm07 (mon.0) 214 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:35.051 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:33.723072+0000 mon.vm07 (mon.0) 214 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:35.051 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:33.725353+0000 mon.vm07 (mon.0) 215 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:35.051 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:33.725353+0000 mon.vm07 (mon.0) 215 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:35.051 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:33.727441+0000 mon.vm07 (mon.0) 216 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:35.051 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:33.727441+0000 mon.vm07 (mon.0) 216 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: cephadm 2026-03-07T06:54:33.728364+0000 mgr.vm07.yrfcuj (mgr.14201) 16 : cephadm [INF] Deploying daemon node-exporter.vm09 on vm09 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: cephadm 2026-03-07T06:54:33.728364+0000 mgr.vm07.yrfcuj (mgr.14201) 16 : cephadm [INF] Deploying daemon node-exporter.vm09 on vm09 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:34.433757+0000 mon.vm07 (mon.0) 217 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:34.433757+0000 mon.vm07 (mon.0) 217 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:34.436089+0000 mon.vm07 (mon.0) 218 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:34.436089+0000 mon.vm07 (mon.0) 218 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:34.438285+0000 mon.vm07 (mon.0) 219 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:34.438285+0000 mon.vm07 (mon.0) 219 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:34.440007+0000 mon.vm07 (mon.0) 220 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:34.440007+0000 mon.vm07 (mon.0) 220 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:34.441231+0000 mon.vm07 (mon.0) 221 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm09.eqznpw", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:34.441231+0000 mon.vm07 (mon.0) 221 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm09.eqznpw", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:34.442162+0000 mon.vm07 (mon.0) 222 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.vm09.eqznpw", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:34.442162+0000 mon.vm07 (mon.0) 222 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.vm09.eqznpw", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:34.443117+0000 mon.vm07 (mon.0) 223 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:34.443117+0000 mon.vm07 (mon.0) 223 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:34.443571+0000 mon.vm07 (mon.0) 224 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: audit 2026-03-07T06:54:34.443571+0000 mon.vm07 (mon.0) 224 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: cephadm 2026-03-07T06:54:34.444007+0000 mgr.vm07.yrfcuj (mgr.14201) 17 : cephadm [INF] Deploying daemon mgr.vm09.eqznpw on vm09 2026-03-07T07:54:35.052 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:34 vm07 bash[17031]: cephadm 2026-03-07T06:54:34.444007+0000 mgr.vm07.yrfcuj (mgr.14201) 17 : cephadm [INF] Deploying daemon mgr.vm09.eqznpw on vm09 2026-03-07T07:54:36.328 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 systemd[1]: Started Ceph mon.vm09 for 312fdbc4-19f2-11f1-81d0-4bbd10a1e012. 2026-03-07T07:54:36.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:36 vm07 bash[17031]: audit 2026-03-07T06:54:35.108316+0000 mon.vm07 (mon.0) 225 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:36 vm07 bash[17031]: audit 2026-03-07T06:54:35.108316+0000 mon.vm07 (mon.0) 225 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:36 vm07 bash[17031]: audit 2026-03-07T06:54:35.171990+0000 mon.vm07 (mon.0) 226 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:36 vm07 bash[17031]: audit 2026-03-07T06:54:35.171990+0000 mon.vm07 (mon.0) 226 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:36 vm07 bash[17031]: audit 2026-03-07T06:54:35.174642+0000 mon.vm07 (mon.0) 227 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:36 vm07 bash[17031]: audit 2026-03-07T06:54:35.174642+0000 mon.vm07 (mon.0) 227 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:36 vm07 bash[17031]: audit 2026-03-07T06:54:35.182423+0000 mon.vm07 (mon.0) 228 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:36 vm07 bash[17031]: audit 2026-03-07T06:54:35.182423+0000 mon.vm07 (mon.0) 228 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:36 vm07 bash[17031]: audit 2026-03-07T06:54:35.185135+0000 mon.vm07 (mon.0) 229 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:36 vm07 bash[17031]: audit 2026-03-07T06:54:35.185135+0000 mon.vm07 (mon.0) 229 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:36 vm07 bash[17031]: audit 2026-03-07T06:54:35.186291+0000 mon.vm07 (mon.0) 230 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T07:54:36.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:36 vm07 bash[17031]: audit 2026-03-07T06:54:35.186291+0000 mon.vm07 (mon.0) 230 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T07:54:36.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:36 vm07 bash[17031]: audit 2026-03-07T06:54:35.186760+0000 mon.vm07 (mon.0) 231 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:36.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:36 vm07 bash[17031]: audit 2026-03-07T06:54:35.186760+0000 mon.vm07 (mon.0) 231 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:36.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:36 vm07 bash[17031]: cephadm 2026-03-07T06:54:35.187359+0000 mgr.vm07.yrfcuj (mgr.14201) 18 : cephadm [INF] Deploying daemon mon.vm09 on vm09 2026-03-07T07:54:36.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:36 vm07 bash[17031]: cephadm 2026-03-07T06:54:35.187359+0000 mgr.vm07.yrfcuj (mgr.14201) 18 : cephadm [INF] Deploying daemon mon.vm09 on vm09 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.526+0000 7facd4c6ad80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.526+0000 7facd4c6ad80 0 ceph version 19.2.3-39-g340d3c24fc6 (340d3c24fc6ae7529322dc7ccee6c6cb2589da0a) squid (stable), process ceph-mon, pid 7 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.526+0000 7facd4c6ad80 0 pidfile_write: ignore empty --pid-file 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 0 load: jerasure load: lrc 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: RocksDB version: 7.9.2 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Git sha 0 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Compile date 2026-03-06 13:52:12 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: DB SUMMARY 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: DB Session ID: VFI1DU7C1L2D2O3IEFIY 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: CURRENT file: CURRENT 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: MANIFEST file: MANIFEST-000005 size: 59 Bytes 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-vm09/store.db dir, Total Num: 0, files: 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-vm09/store.db: 000004.log size: 511 ; 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.error_if_exists: 0 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.create_if_missing: 0 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.paranoid_checks: 1 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.env: 0x5611b3fb5ca0 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.info_log: 0x5611f27c9820 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.statistics: (nil) 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.use_fsync: 0 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.max_log_file_size: 0 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-07T07:54:36.578 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.allow_fallocate: 1 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.use_direct_reads: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.db_log_dir: 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.wal_dir: 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.write_buffer_manager: 0x5611f27cd900 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.unordered_write: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.row_cache: None 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.wal_filter: None 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.two_write_queues: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.wal_compression: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.atomic_flush: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.log_readahead_size: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-07T07:54:36.579 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.max_background_jobs: 2 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.max_background_compactions: -1 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.max_subcompactions: 1 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.max_open_files: -1 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.max_background_flushes: -1 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Compression algorithms supported: 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: kZSTD supported: 0 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: kXpressCompression supported: 0 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: kBZip2Compression supported: 0 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: kLZ4Compression supported: 1 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: kZlibCompression supported: 1 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: kSnappyCompression supported: 1 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-vm09/store.db/MANIFEST-000005 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.merge_operator: 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.compaction_filter: None 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x5611f27c9460) 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cache_index_and_filter_blocks: 1 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: pin_top_level_index_and_filter: 1 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: index_type: 0 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: data_block_index_type: 0 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: index_shortening: 1 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: data_block_hash_table_util_ratio: 0.750000 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: checksum: 4 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: no_block_cache: 0 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: block_cache: 0x5611f27ef350 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: block_cache_name: BinnedLRUCache 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: block_cache_options: 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: capacity : 536870912 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: num_shard_bits : 4 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: strict_capacity_limit : 0 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: high_pri_pool_ratio: 0.000 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: block_cache_compressed: (nil) 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: persistent_cache: (nil) 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: block_size: 4096 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: block_size_deviation: 10 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: block_restart_interval: 16 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: index_block_restart_interval: 1 2026-03-07T07:54:36.580 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: metadata_block_size: 4096 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: partition_filters: 0 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: use_delta_encoding: 1 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: filter_policy: bloomfilter 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: whole_key_filtering: 1 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: verify_compression: 0 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: read_amp_bytes_per_bit: 0 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: format_version: 5 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: enable_index_compression: 1 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: block_align: 0 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: max_auto_readahead_size: 262144 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: prepopulate_block_cache: 0 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: initial_auto_readahead_size: 8192 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: num_file_reads_for_auto_readahead: 2 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.compression: NoCompression 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.num_levels: 7 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.530+0000 7facd4c6ad80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-07T07:54:36.581 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.inplace_update_support: 0 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.bloom_locality: 0 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.max_successive_merges: 0 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.ttl: 2592000 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.enable_blob_files: false 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.min_blob_size: 0 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-vm09/store.db/MANIFEST-000005 succeeded,manifest_file_number is 5, next_file_number is 7, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: 3843a162-e18c-4750-b907-23b72a168397 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772866476535974, "job": 1, "event": "recovery_started", "wal_files": [4]} 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.534+0000 7facd4c6ad80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #4 mode 2 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.542+0000 7facd4c6ad80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772866476543646, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 8, "file_size": 1643, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 1, "largest_seqno": 5, "table_properties": {"data_size": 523, "index_size": 31, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 115, "raw_average_key_size": 23, "raw_value_size": 401, "raw_average_value_size": 80, "num_data_blocks": 1, "num_entries": 5, "num_filter_entries": 5, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1772866476, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "3843a162-e18c-4750-b907-23b72a168397", "db_session_id": "VFI1DU7C1L2D2O3IEFIY", "orig_file_number": 8, "seqno_to_time_mapping": "N/A"}} 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.542+0000 7facd4c6ad80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1772866476543700, "job": 1, "event": "recovery_finished"} 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.542+0000 7facd4c6ad80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 10 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.546+0000 7facd4c6ad80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-vm09/store.db/000004.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.546+0000 7facd4c6ad80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x5611f27f0e00 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.546+0000 7facd4c6ad80 4 rocksdb: DB pointer 0x5611f2904000 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.546+0000 7facd4c6ad80 0 mon.vm09 does not exist in monmap, will attempt to join an existing cluster 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.546+0000 7facd4c6ad80 0 using public_addr v2:192.168.123.109:0/0 -> [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.546+0000 7facd4c6ad80 0 starting mon.vm09 rank -1 at public addrs [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] at bind addrs [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon_data /var/lib/ceph/mon/ceph-vm09 fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.546+0000 7facd4c6ad80 1 mon.vm09@-1(???) e0 preinit fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.546+0000 7faccaa34640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.546+0000 7faccaa34640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: ** DB Stats ** 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-07T07:54:36.582 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: ** Compaction Stats [default] ** 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: L0 1/0 1.60 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.2 0.01 0.00 1 0.008 0 0 0.0 0.0 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: Sum 1/0 1.60 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.2 0.01 0.00 1 0.008 0 0 0.0 0.0 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 0.2 0.01 0.00 1 0.008 0 0 0.0 0.0 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: ** Compaction Stats [default] ** 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.2 0.01 0.00 1 0.008 0 0 0.0 0.0 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: Flush(GB): cumulative 0.000, interval 0.000 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: AddFile(Total Files): cumulative 0, interval 0 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: AddFile(Keys): cumulative 0, interval 0 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: Cumulative compaction: 0.00 GB write, 0.11 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: Interval compaction: 0.00 GB write, 0.11 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: Block cache BinnedLRUCache@0x5611f27ef350#7 capacity: 512.00 MB usage: 0.86 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 4e-06 secs_since: 0 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: Block cache entry stats(count,size,portion): DataBlock(1,0.64 KB,0.00012219%) FilterBlock(1,0.11 KB,2.08616e-05%) IndexBlock(1,0.11 KB,2.08616e-05%) Misc(1,0.00 KB,0%) 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: ** File Read Latency Histogram By Level [default] ** 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.570+0000 7faccda3a640 0 mon.vm09@-1(synchronizing).mds e1 new map 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.570+0000 7faccda3a640 0 mon.vm09@-1(synchronizing).mds e1 print_map 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: e1 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: btime 2026-03-07T06:52:52:038351+0000 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: legacy client fscid: -1 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: No filesystems configured 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.574+0000 7faccda3a640 1 mon.vm09@-1(synchronizing).osd e0 _set_cache_ratios kv ratio 0.25 inc ratio 0.375 full ratio 0.375 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.574+0000 7faccda3a640 1 mon.vm09@-1(synchronizing).osd e0 register_cache_with_pcm pcm target: 2147483648 pcm max: 1020054732 pcm min: 134217728 inc_osd_cache size: 1 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.574+0000 7faccda3a640 1 mon.vm09@-1(synchronizing).osd e1 e1: 0 total, 0 up, 0 in 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.574+0000 7faccda3a640 1 mon.vm09@-1(synchronizing).osd e2 e2: 0 total, 0 up, 0 in 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.574+0000 7faccda3a640 1 mon.vm09@-1(synchronizing).osd e3 e3: 0 total, 0 up, 0 in 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.574+0000 7faccda3a640 1 mon.vm09@-1(synchronizing).osd e4 e4: 0 total, 0 up, 0 in 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.574+0000 7faccda3a640 1 mon.vm09@-1(synchronizing).osd e5 e5: 0 total, 0 up, 0 in 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.574+0000 7faccda3a640 0 mon.vm09@-1(synchronizing).osd e5 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.574+0000 7faccda3a640 0 mon.vm09@-1(synchronizing).osd e5 crush map has features 288514050185494528, adjusting msgr requires 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.574+0000 7faccda3a640 0 mon.vm09@-1(synchronizing).osd e5 crush map has features 288514050185494528, adjusting msgr requires 2026-03-07T07:54:36.583 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.574+0000 7faccda3a640 0 mon.vm09@-1(synchronizing).osd e5 crush map has features 288514050185494528, adjusting msgr requires 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:25.130902+0000 mon.vm07 (mon.0) 176 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:25.130902+0000 mon.vm07 (mon.0) 176 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:25.131864+0000 mon.vm07 (mon.0) 177 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/mirror_snapshot_schedule"}]: dispatch 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:25.131864+0000 mon.vm07 (mon.0) 177 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/mirror_snapshot_schedule"}]: dispatch 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:25.167407+0000 mon.vm07 (mon.0) 178 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/trash_purge_schedule"}]: dispatch 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:25.167407+0000 mon.vm07 (mon.0) 178 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm07.yrfcuj/trash_purge_schedule"}]: dispatch 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:25.739320+0000 mon.vm07 (mon.0) 179 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:25.739320+0000 mon.vm07 (mon.0) 179 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cluster 2026-03-07T06:54:26.075119+0000 mon.vm07 (mon.0) 180 : cluster [DBG] mgrmap e16: vm07.yrfcuj(active, since 1.00839s) 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cluster 2026-03-07T06:54:26.075119+0000 mon.vm07 (mon.0) 180 : cluster [DBG] mgrmap e16: vm07.yrfcuj(active, since 1.00839s) 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:25.998985+0000 mgr.vm07.yrfcuj (mgr.14201) 1 : cephadm [INF] [07/Mar/2026:06:54:25] ENGINE Bus STARTING 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:25.998985+0000 mgr.vm07.yrfcuj (mgr.14201) 1 : cephadm [INF] [07/Mar/2026:06:54:25] ENGINE Bus STARTING 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:26.101321+0000 mgr.vm07.yrfcuj (mgr.14201) 2 : cephadm [INF] [07/Mar/2026:06:54:26] ENGINE Serving on http://192.168.123.107:8765 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:26.101321+0000 mgr.vm07.yrfcuj (mgr.14201) 2 : cephadm [INF] [07/Mar/2026:06:54:26] ENGINE Serving on http://192.168.123.107:8765 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:26.210153+0000 mgr.vm07.yrfcuj (mgr.14201) 3 : cephadm [INF] [07/Mar/2026:06:54:26] ENGINE Serving on https://192.168.123.107:7150 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:26.210153+0000 mgr.vm07.yrfcuj (mgr.14201) 3 : cephadm [INF] [07/Mar/2026:06:54:26] ENGINE Serving on https://192.168.123.107:7150 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:26.210271+0000 mgr.vm07.yrfcuj (mgr.14201) 4 : cephadm [INF] [07/Mar/2026:06:54:26] ENGINE Bus STARTED 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:26.210271+0000 mgr.vm07.yrfcuj (mgr.14201) 4 : cephadm [INF] [07/Mar/2026:06:54:26] ENGINE Bus STARTED 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:26.210489+0000 mgr.vm07.yrfcuj (mgr.14201) 5 : cephadm [INF] [07/Mar/2026:06:54:26] ENGINE Client ('192.168.123.107', 47182) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:26.210489+0000 mgr.vm07.yrfcuj (mgr.14201) 5 : cephadm [INF] [07/Mar/2026:06:54:26] ENGINE Client ('192.168.123.107', 47182) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:27.022992+0000 mon.vm07 (mon.0) 181 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:27.022992+0000 mon.vm07 (mon.0) 181 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:27.615589+0000 mon.vm07 (mon.0) 182 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:27.615589+0000 mon.vm07 (mon.0) 182 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:27.745479+0000 mon.vm07 (mon.0) 183 : audit [DBG] from='client.? 192.168.123.109:0/1396054845' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:27.745479+0000 mon.vm07 (mon.0) 183 : audit [DBG] from='client.? 192.168.123.109:0/1396054845' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cluster 2026-03-07T06:54:28.027190+0000 mon.vm07 (mon.0) 184 : cluster [DBG] mgrmap e17: vm07.yrfcuj(active, since 2s) 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cluster 2026-03-07T06:54:28.027190+0000 mon.vm07 (mon.0) 184 : cluster [DBG] mgrmap e17: vm07.yrfcuj(active, since 2s) 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:30.668232+0000 mon.vm07 (mon.0) 185 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:30.668232+0000 mon.vm07 (mon.0) 185 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:30.670781+0000 mon.vm07 (mon.0) 186 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.851 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:30.670781+0000 mon.vm07 (mon.0) 186 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:30.673303+0000 mon.vm07 (mon.0) 187 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:30.673303+0000 mon.vm07 (mon.0) 187 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:30.675499+0000 mon.vm07 (mon.0) 188 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:30.675499+0000 mon.vm07 (mon.0) 188 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:30.676021+0000 mon.vm07 (mon.0) 189 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:30.676021+0000 mon.vm07 (mon.0) 189 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:30.972282+0000 mon.vm07 (mon.0) 190 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:30.972282+0000 mon.vm07 (mon.0) 190 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:30.974998+0000 mon.vm07 (mon.0) 191 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:30.974998+0000 mon.vm07 (mon.0) 191 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.488452+0000 mon.vm07 (mon.0) 192 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.488452+0000 mon.vm07 (mon.0) 192 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.490892+0000 mon.vm07 (mon.0) 193 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.490892+0000 mon.vm07 (mon.0) 193 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.491488+0000 mon.vm07 (mon.0) 194 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"}]: dispatch 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.491488+0000 mon.vm07 (mon.0) 194 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"}]: dispatch 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.492035+0000 mon.vm07 (mon.0) 195 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.492035+0000 mon.vm07 (mon.0) 195 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.492359+0000 mon.vm07 (mon.0) 196 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.492359+0000 mon.vm07 (mon.0) 196 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.625758+0000 mon.vm07 (mon.0) 197 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.625758+0000 mon.vm07 (mon.0) 197 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.627957+0000 mon.vm07 (mon.0) 198 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.627957+0000 mon.vm07 (mon.0) 198 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.629725+0000 mon.vm07 (mon.0) 199 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.629725+0000 mon.vm07 (mon.0) 199 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.631776+0000 mon.vm07 (mon.0) 200 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.631776+0000 mon.vm07 (mon.0) 200 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.633777+0000 mon.vm07 (mon.0) 201 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.633777+0000 mon.vm07 (mon.0) 201 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.634569+0000 mon.vm07 (mon.0) 202 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.634569+0000 mon.vm07 (mon.0) 202 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.635654+0000 mon.vm07 (mon.0) 203 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.635654+0000 mon.vm07 (mon.0) 203 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.636763+0000 mon.vm07 (mon.0) 204 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:31.636763+0000 mon.vm07 (mon.0) 204 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:31.492906+0000 mgr.vm07.yrfcuj (mgr.14201) 6 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:31.492906+0000 mgr.vm07.yrfcuj (mgr.14201) 6 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:31.492986+0000 mgr.vm07.yrfcuj (mgr.14201) 7 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:31.492986+0000 mgr.vm07.yrfcuj (mgr.14201) 7 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-07T07:54:36.852 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:31.525712+0000 mgr.vm07.yrfcuj (mgr.14201) 8 : cephadm [INF] Updating vm09:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.conf 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:31.525712+0000 mgr.vm07.yrfcuj (mgr.14201) 8 : cephadm [INF] Updating vm09:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.conf 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:31.525781+0000 mgr.vm07.yrfcuj (mgr.14201) 9 : cephadm [INF] Updating vm07:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.conf 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:31.525781+0000 mgr.vm07.yrfcuj (mgr.14201) 9 : cephadm [INF] Updating vm07:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.conf 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:31.557402+0000 mgr.vm07.yrfcuj (mgr.14201) 10 : cephadm [INF] Updating vm07:/etc/ceph/ceph.client.admin.keyring 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:31.557402+0000 mgr.vm07.yrfcuj (mgr.14201) 10 : cephadm [INF] Updating vm07:/etc/ceph/ceph.client.admin.keyring 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:31.557648+0000 mgr.vm07.yrfcuj (mgr.14201) 11 : cephadm [INF] Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:31.557648+0000 mgr.vm07.yrfcuj (mgr.14201) 11 : cephadm [INF] Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:31.589971+0000 mgr.vm07.yrfcuj (mgr.14201) 12 : cephadm [INF] Updating vm07:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.client.admin.keyring 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:31.589971+0000 mgr.vm07.yrfcuj (mgr.14201) 12 : cephadm [INF] Updating vm07:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.client.admin.keyring 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:31.590191+0000 mgr.vm07.yrfcuj (mgr.14201) 13 : cephadm [INF] Updating vm09:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.client.admin.keyring 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:31.590191+0000 mgr.vm07.yrfcuj (mgr.14201) 13 : cephadm [INF] Updating vm09:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.client.admin.keyring 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:31.637330+0000 mgr.vm07.yrfcuj (mgr.14201) 14 : cephadm [INF] Deploying daemon ceph-exporter.vm09 on vm09 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:31.637330+0000 mgr.vm07.yrfcuj (mgr.14201) 14 : cephadm [INF] Deploying daemon ceph-exporter.vm09 on vm09 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:32.811912+0000 mon.vm07 (mon.0) 205 : audit [DBG] from='client.? 192.168.123.109:0/1490086917' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:32.811912+0000 mon.vm07 (mon.0) 205 : audit [DBG] from='client.? 192.168.123.109:0/1490086917' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:32.920774+0000 mon.vm07 (mon.0) 206 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:32.920774+0000 mon.vm07 (mon.0) 206 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:32.923304+0000 mon.vm07 (mon.0) 207 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:32.923304+0000 mon.vm07 (mon.0) 207 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:32.925626+0000 mon.vm07 (mon.0) 208 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:32.925626+0000 mon.vm07 (mon.0) 208 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:32.927598+0000 mon.vm07 (mon.0) 209 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:32.927598+0000 mon.vm07 (mon.0) 209 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:32.928329+0000 mon.vm07 (mon.0) 210 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:32.928329+0000 mon.vm07 (mon.0) 210 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:32.929293+0000 mon.vm07 (mon.0) 211 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-07T07:54:36.853 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:32.929293+0000 mon.vm07 (mon.0) 211 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:32.930583+0000 mon.vm07 (mon.0) 212 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:32.930583+0000 mon.vm07 (mon.0) 212 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:32.931049+0000 mgr.vm07.yrfcuj (mgr.14201) 15 : cephadm [INF] Deploying daemon crash.vm09 on vm09 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:32.931049+0000 mgr.vm07.yrfcuj (mgr.14201) 15 : cephadm [INF] Deploying daemon crash.vm09 on vm09 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:33.720078+0000 mon.vm07 (mon.0) 213 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:33.720078+0000 mon.vm07 (mon.0) 213 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:33.723072+0000 mon.vm07 (mon.0) 214 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:33.723072+0000 mon.vm07 (mon.0) 214 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:33.725353+0000 mon.vm07 (mon.0) 215 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:33.725353+0000 mon.vm07 (mon.0) 215 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:33.727441+0000 mon.vm07 (mon.0) 216 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:33.727441+0000 mon.vm07 (mon.0) 216 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:33.728364+0000 mgr.vm07.yrfcuj (mgr.14201) 16 : cephadm [INF] Deploying daemon node-exporter.vm09 on vm09 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:33.728364+0000 mgr.vm07.yrfcuj (mgr.14201) 16 : cephadm [INF] Deploying daemon node-exporter.vm09 on vm09 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:34.433757+0000 mon.vm07 (mon.0) 217 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:34.433757+0000 mon.vm07 (mon.0) 217 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:34.436089+0000 mon.vm07 (mon.0) 218 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:34.436089+0000 mon.vm07 (mon.0) 218 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:34.438285+0000 mon.vm07 (mon.0) 219 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:34.438285+0000 mon.vm07 (mon.0) 219 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:34.440007+0000 mon.vm07 (mon.0) 220 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:34.440007+0000 mon.vm07 (mon.0) 220 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:34.441231+0000 mon.vm07 (mon.0) 221 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm09.eqznpw", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:34.441231+0000 mon.vm07 (mon.0) 221 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm09.eqznpw", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:34.442162+0000 mon.vm07 (mon.0) 222 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.vm09.eqznpw", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:34.442162+0000 mon.vm07 (mon.0) 222 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.vm09.eqznpw", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:34.443117+0000 mon.vm07 (mon.0) 223 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:34.443117+0000 mon.vm07 (mon.0) 223 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:34.443571+0000 mon.vm07 (mon.0) 224 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:34.443571+0000 mon.vm07 (mon.0) 224 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:34.444007+0000 mgr.vm07.yrfcuj (mgr.14201) 17 : cephadm [INF] Deploying daemon mgr.vm09.eqznpw on vm09 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:34.444007+0000 mgr.vm07.yrfcuj (mgr.14201) 17 : cephadm [INF] Deploying daemon mgr.vm09.eqznpw on vm09 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:35.108316+0000 mon.vm07 (mon.0) 225 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:35.108316+0000 mon.vm07 (mon.0) 225 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:35.171990+0000 mon.vm07 (mon.0) 226 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:35.171990+0000 mon.vm07 (mon.0) 226 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:35.174642+0000 mon.vm07 (mon.0) 227 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:35.174642+0000 mon.vm07 (mon.0) 227 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:35.182423+0000 mon.vm07 (mon.0) 228 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:35.182423+0000 mon.vm07 (mon.0) 228 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.854 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:35.185135+0000 mon.vm07 (mon.0) 229 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.855 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:35.185135+0000 mon.vm07 (mon.0) 229 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:36.855 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:35.186291+0000 mon.vm07 (mon.0) 230 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T07:54:36.855 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:35.186291+0000 mon.vm07 (mon.0) 230 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T07:54:36.855 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:35.186760+0000 mon.vm07 (mon.0) 231 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:36.855 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: audit 2026-03-07T06:54:35.186760+0000 mon.vm07 (mon.0) 231 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:36.855 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:35.187359+0000 mgr.vm07.yrfcuj (mgr.14201) 18 : cephadm [INF] Deploying daemon mon.vm09 on vm09 2026-03-07T07:54:36.855 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: cephadm 2026-03-07T06:54:35.187359+0000 mgr.vm07.yrfcuj (mgr.14201) 18 : cephadm [INF] Deploying daemon mon.vm09 on vm09 2026-03-07T07:54:36.855 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:36 vm09 bash[21220]: debug 2026-03-07T06:54:36.606+0000 7faccda3a640 1 mon.vm09@-1(synchronizing).paxosservice(auth 1..8) refresh upgraded, format 0 -> 3 2026-03-07T07:54:40.803 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm09/config 2026-03-07T07:54:42.019 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:36.613218+0000 mon.vm07 (mon.0) 239 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:54:42.019 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:36.613218+0000 mon.vm07 (mon.0) 239 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:54:42.019 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:36.613298+0000 mon.vm07 (mon.0) 240 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.019 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:36.613298+0000 mon.vm07 (mon.0) 240 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.019 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:36.613536+0000 mon.vm07 (mon.0) 241 : cluster [INF] mon.vm07 calling monitor election 2026-03-07T07:54:42.019 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:36.613536+0000 mon.vm07 (mon.0) 241 : cluster [INF] mon.vm07 calling monitor election 2026-03-07T07:54:42.019 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:37.609506+0000 mon.vm07 (mon.0) 242 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.019 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:37.609506+0000 mon.vm07 (mon.0) 242 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.019 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:38.609691+0000 mon.vm07 (mon.0) 243 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.019 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:38.609691+0000 mon.vm07 (mon.0) 243 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.019 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:38.616748+0000 mon.vm09 (mon.1) 1 : cluster [INF] mon.vm09 calling monitor election 2026-03-07T07:54:42.019 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:38.616748+0000 mon.vm09 (mon.1) 1 : cluster [INF] mon.vm09 calling monitor election 2026-03-07T07:54:42.019 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:39.609677+0000 mon.vm07 (mon.0) 244 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.019 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:39.609677+0000 mon.vm07 (mon.0) 244 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.019 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:40.124027+0000 mon.vm07 (mon.0) 245 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:54:42.019 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:40.124027+0000 mon.vm07 (mon.0) 245 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:54:42.019 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:40.609740+0000 mon.vm07 (mon.0) 246 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.019 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:40.609740+0000 mon.vm07 (mon.0) 246 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:41.609898+0000 mon.vm07 (mon.0) 247 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:41.609898+0000 mon.vm07 (mon.0) 247 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.618887+0000 mon.vm07 (mon.0) 248 : cluster [INF] mon.vm07 is new leader, mons vm07,vm09 in quorum (ranks 0,1) 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.618887+0000 mon.vm07 (mon.0) 248 : cluster [INF] mon.vm07 is new leader, mons vm07,vm09 in quorum (ranks 0,1) 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622480+0000 mon.vm07 (mon.0) 249 : cluster [DBG] monmap epoch 2 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622480+0000 mon.vm07 (mon.0) 249 : cluster [DBG] monmap epoch 2 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622500+0000 mon.vm07 (mon.0) 250 : cluster [DBG] fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622500+0000 mon.vm07 (mon.0) 250 : cluster [DBG] fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622509+0000 mon.vm07 (mon.0) 251 : cluster [DBG] last_changed 2026-03-07T06:54:36.610505+0000 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622509+0000 mon.vm07 (mon.0) 251 : cluster [DBG] last_changed 2026-03-07T06:54:36.610505+0000 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622517+0000 mon.vm07 (mon.0) 252 : cluster [DBG] created 2026-03-07T06:52:50.872627+0000 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622517+0000 mon.vm07 (mon.0) 252 : cluster [DBG] created 2026-03-07T06:52:50.872627+0000 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622527+0000 mon.vm07 (mon.0) 253 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622527+0000 mon.vm07 (mon.0) 253 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622536+0000 mon.vm07 (mon.0) 254 : cluster [DBG] election_strategy: 1 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622536+0000 mon.vm07 (mon.0) 254 : cluster [DBG] election_strategy: 1 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622546+0000 mon.vm07 (mon.0) 255 : cluster [DBG] 0: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.vm07 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622546+0000 mon.vm07 (mon.0) 255 : cluster [DBG] 0: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.vm07 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622555+0000 mon.vm07 (mon.0) 256 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.vm09 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622555+0000 mon.vm07 (mon.0) 256 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.vm09 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622894+0000 mon.vm07 (mon.0) 257 : cluster [DBG] fsmap 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622894+0000 mon.vm07 (mon.0) 257 : cluster [DBG] fsmap 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622916+0000 mon.vm07 (mon.0) 258 : cluster [DBG] osdmap e5: 0 total, 0 up, 0 in 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.622916+0000 mon.vm07 (mon.0) 258 : cluster [DBG] osdmap e5: 0 total, 0 up, 0 in 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.623022+0000 mon.vm07 (mon.0) 259 : cluster [DBG] mgrmap e17: vm07.yrfcuj(active, since 16s) 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.623022+0000 mon.vm07 (mon.0) 259 : cluster [DBG] mgrmap e17: vm07.yrfcuj(active, since 16s) 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.623098+0000 mon.vm07 (mon.0) 260 : cluster [INF] overall HEALTH_OK 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: cluster 2026-03-07T06:54:41.623098+0000 mon.vm07 (mon.0) 260 : cluster [INF] overall HEALTH_OK 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:41.626757+0000 mon.vm07 (mon.0) 261 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:41.626757+0000 mon.vm07 (mon.0) 261 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:41.629712+0000 mon.vm07 (mon.0) 262 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:41.629712+0000 mon.vm07 (mon.0) 262 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:41.632908+0000 mon.vm07 (mon.0) 263 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:41.632908+0000 mon.vm07 (mon.0) 263 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:41.633744+0000 mon.vm07 (mon.0) 264 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:41.633744+0000 mon.vm07 (mon.0) 264 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:41.634296+0000 mon.vm07 (mon.0) 265 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:41 vm07 bash[17031]: audit 2026-03-07T06:54:41.634296+0000 mon.vm07 (mon.0) 265 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:36.613218+0000 mon.vm07 (mon.0) 239 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:36.613218+0000 mon.vm07 (mon.0) 239 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:36.613298+0000 mon.vm07 (mon.0) 240 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:36.613298+0000 mon.vm07 (mon.0) 240 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.020 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:36.613536+0000 mon.vm07 (mon.0) 241 : cluster [INF] mon.vm07 calling monitor election 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:36.613536+0000 mon.vm07 (mon.0) 241 : cluster [INF] mon.vm07 calling monitor election 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:37.609506+0000 mon.vm07 (mon.0) 242 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:37.609506+0000 mon.vm07 (mon.0) 242 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:38.609691+0000 mon.vm07 (mon.0) 243 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:38.609691+0000 mon.vm07 (mon.0) 243 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:38.616748+0000 mon.vm09 (mon.1) 1 : cluster [INF] mon.vm09 calling monitor election 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:38.616748+0000 mon.vm09 (mon.1) 1 : cluster [INF] mon.vm09 calling monitor election 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:39.609677+0000 mon.vm07 (mon.0) 244 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:39.609677+0000 mon.vm07 (mon.0) 244 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:40.124027+0000 mon.vm07 (mon.0) 245 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:40.124027+0000 mon.vm07 (mon.0) 245 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:40.609740+0000 mon.vm07 (mon.0) 246 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:40.609740+0000 mon.vm07 (mon.0) 246 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:41.609898+0000 mon.vm07 (mon.0) 247 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:41.609898+0000 mon.vm07 (mon.0) 247 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.618887+0000 mon.vm07 (mon.0) 248 : cluster [INF] mon.vm07 is new leader, mons vm07,vm09 in quorum (ranks 0,1) 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.618887+0000 mon.vm07 (mon.0) 248 : cluster [INF] mon.vm07 is new leader, mons vm07,vm09 in quorum (ranks 0,1) 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622480+0000 mon.vm07 (mon.0) 249 : cluster [DBG] monmap epoch 2 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622480+0000 mon.vm07 (mon.0) 249 : cluster [DBG] monmap epoch 2 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622500+0000 mon.vm07 (mon.0) 250 : cluster [DBG] fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622500+0000 mon.vm07 (mon.0) 250 : cluster [DBG] fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622509+0000 mon.vm07 (mon.0) 251 : cluster [DBG] last_changed 2026-03-07T06:54:36.610505+0000 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622509+0000 mon.vm07 (mon.0) 251 : cluster [DBG] last_changed 2026-03-07T06:54:36.610505+0000 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622517+0000 mon.vm07 (mon.0) 252 : cluster [DBG] created 2026-03-07T06:52:50.872627+0000 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622517+0000 mon.vm07 (mon.0) 252 : cluster [DBG] created 2026-03-07T06:52:50.872627+0000 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622527+0000 mon.vm07 (mon.0) 253 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622527+0000 mon.vm07 (mon.0) 253 : cluster [DBG] min_mon_release 19 (squid) 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622536+0000 mon.vm07 (mon.0) 254 : cluster [DBG] election_strategy: 1 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622536+0000 mon.vm07 (mon.0) 254 : cluster [DBG] election_strategy: 1 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622546+0000 mon.vm07 (mon.0) 255 : cluster [DBG] 0: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.vm07 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622546+0000 mon.vm07 (mon.0) 255 : cluster [DBG] 0: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.vm07 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622555+0000 mon.vm07 (mon.0) 256 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.vm09 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622555+0000 mon.vm07 (mon.0) 256 : cluster [DBG] 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.vm09 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622894+0000 mon.vm07 (mon.0) 257 : cluster [DBG] fsmap 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622894+0000 mon.vm07 (mon.0) 257 : cluster [DBG] fsmap 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622916+0000 mon.vm07 (mon.0) 258 : cluster [DBG] osdmap e5: 0 total, 0 up, 0 in 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.622916+0000 mon.vm07 (mon.0) 258 : cluster [DBG] osdmap e5: 0 total, 0 up, 0 in 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.623022+0000 mon.vm07 (mon.0) 259 : cluster [DBG] mgrmap e17: vm07.yrfcuj(active, since 16s) 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.623022+0000 mon.vm07 (mon.0) 259 : cluster [DBG] mgrmap e17: vm07.yrfcuj(active, since 16s) 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.623098+0000 mon.vm07 (mon.0) 260 : cluster [INF] overall HEALTH_OK 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: cluster 2026-03-07T06:54:41.623098+0000 mon.vm07 (mon.0) 260 : cluster [INF] overall HEALTH_OK 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:41.626757+0000 mon.vm07 (mon.0) 261 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:41.626757+0000 mon.vm07 (mon.0) 261 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:41.629712+0000 mon.vm07 (mon.0) 262 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:41.629712+0000 mon.vm07 (mon.0) 262 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:41.632908+0000 mon.vm07 (mon.0) 263 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:41.632908+0000 mon.vm07 (mon.0) 263 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:41.633744+0000 mon.vm07 (mon.0) 264 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:41.633744+0000 mon.vm07 (mon.0) 264 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:41.634296+0000 mon.vm07 (mon.0) 265 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:54:42.021 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:41 vm09 bash[21220]: audit 2026-03-07T06:54:41.634296+0000 mon.vm07 (mon.0) 265 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:54:42.087 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-07T07:54:42.087 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":2,"fsid":"312fdbc4-19f2-11f1-81d0-4bbd10a1e012","modified":"2026-03-07T06:54:36.610505Z","created":"2026-03-07T06:52:50.872627Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm07","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:3300","nonce":0},{"type":"v1","addr":"192.168.123.107:6789","nonce":0}]},"addr":"192.168.123.107:6789/0","public_addr":"192.168.123.107:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"vm09","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:3300","nonce":0},{"type":"v1","addr":"192.168.123.109:6789","nonce":0}]},"addr":"192.168.123.109:6789/0","public_addr":"192.168.123.109:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-07T07:54:42.088 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 2 2026-03-07T07:54:42.203 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-07T07:54:42.204 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph config generate-minimal-conf 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: cephadm 2026-03-07T06:54:41.634947+0000 mgr.vm07.yrfcuj (mgr.14201) 19 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: cephadm 2026-03-07T06:54:41.634947+0000 mgr.vm07.yrfcuj (mgr.14201) 19 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: cephadm 2026-03-07T06:54:41.635366+0000 mgr.vm07.yrfcuj (mgr.14201) 20 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: cephadm 2026-03-07T06:54:41.635366+0000 mgr.vm07.yrfcuj (mgr.14201) 20 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: cephadm 2026-03-07T06:54:41.666460+0000 mgr.vm07.yrfcuj (mgr.14201) 21 : cephadm [INF] Updating vm07:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.conf 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: cephadm 2026-03-07T06:54:41.666460+0000 mgr.vm07.yrfcuj (mgr.14201) 21 : cephadm [INF] Updating vm07:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.conf 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: cephadm 2026-03-07T06:54:41.668627+0000 mgr.vm07.yrfcuj (mgr.14201) 22 : cephadm [INF] Updating vm09:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.conf 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: cephadm 2026-03-07T06:54:41.668627+0000 mgr.vm07.yrfcuj (mgr.14201) 22 : cephadm [INF] Updating vm09:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.conf 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:41.725113+0000 mon.vm07 (mon.0) 266 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:41.725113+0000 mon.vm07 (mon.0) 266 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:41.728468+0000 mon.vm07 (mon.0) 267 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:41.728468+0000 mon.vm07 (mon.0) 267 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:41.732045+0000 mon.vm07 (mon.0) 268 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:41.732045+0000 mon.vm07 (mon.0) 268 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:41.735227+0000 mon.vm07 (mon.0) 269 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:41.735227+0000 mon.vm07 (mon.0) 269 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:41.738801+0000 mon.vm07 (mon.0) 270 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:41.738801+0000 mon.vm07 (mon.0) 270 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: cephadm 2026-03-07T06:54:41.750369+0000 mgr.vm07.yrfcuj (mgr.14201) 23 : cephadm [INF] Reconfiguring mon.vm07 (unknown last config time)... 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: cephadm 2026-03-07T06:54:41.750369+0000 mgr.vm07.yrfcuj (mgr.14201) 23 : cephadm [INF] Reconfiguring mon.vm07 (unknown last config time)... 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:41.750658+0000 mon.vm07 (mon.0) 271 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:41.750658+0000 mon.vm07 (mon.0) 271 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:41.751103+0000 mon.vm07 (mon.0) 272 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:41.751103+0000 mon.vm07 (mon.0) 272 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:41.751444+0000 mon.vm07 (mon.0) 273 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:41.751444+0000 mon.vm07 (mon.0) 273 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: cephadm 2026-03-07T06:54:41.751835+0000 mgr.vm07.yrfcuj (mgr.14201) 24 : cephadm [INF] Reconfiguring daemon mon.vm07 on vm07 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: cephadm 2026-03-07T06:54:41.751835+0000 mgr.vm07.yrfcuj (mgr.14201) 24 : cephadm [INF] Reconfiguring daemon mon.vm07 on vm07 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:42.081935+0000 mon.vm07 (mon.0) 274 : audit [DBG] from='client.? 192.168.123.109:0/529801458' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:42.081935+0000 mon.vm07 (mon.0) 274 : audit [DBG] from='client.? 192.168.123.109:0/529801458' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:42.152767+0000 mon.vm07 (mon.0) 275 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:42.152767+0000 mon.vm07 (mon.0) 275 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:42.156780+0000 mon.vm07 (mon.0) 276 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:42.156780+0000 mon.vm07 (mon.0) 276 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: cephadm 2026-03-07T06:54:42.157472+0000 mgr.vm07.yrfcuj (mgr.14201) 25 : cephadm [INF] Reconfiguring prometheus.vm07 (dependencies changed)... 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: cephadm 2026-03-07T06:54:42.157472+0000 mgr.vm07.yrfcuj (mgr.14201) 25 : cephadm [INF] Reconfiguring prometheus.vm07 (dependencies changed)... 2026-03-07T07:54:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: cephadm 2026-03-07T06:54:42.323244+0000 mgr.vm07.yrfcuj (mgr.14201) 26 : cephadm [INF] Reconfiguring daemon prometheus.vm07 on vm07 2026-03-07T07:54:42.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: cephadm 2026-03-07T06:54:42.323244+0000 mgr.vm07.yrfcuj (mgr.14201) 26 : cephadm [INF] Reconfiguring daemon prometheus.vm07 on vm07 2026-03-07T07:54:42.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:42.610058+0000 mon.vm07 (mon.0) 277 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:42 vm09 bash[21220]: audit 2026-03-07T06:54:42.610058+0000 mon.vm07 (mon.0) 277 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.920 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: cephadm 2026-03-07T06:54:41.634947+0000 mgr.vm07.yrfcuj (mgr.14201) 19 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-07T07:54:42.920 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: cephadm 2026-03-07T06:54:41.634947+0000 mgr.vm07.yrfcuj (mgr.14201) 19 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-07T07:54:42.920 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: cephadm 2026-03-07T06:54:41.635366+0000 mgr.vm07.yrfcuj (mgr.14201) 20 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-07T07:54:42.920 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: cephadm 2026-03-07T06:54:41.635366+0000 mgr.vm07.yrfcuj (mgr.14201) 20 : cephadm [INF] Updating vm09:/etc/ceph/ceph.conf 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: cephadm 2026-03-07T06:54:41.666460+0000 mgr.vm07.yrfcuj (mgr.14201) 21 : cephadm [INF] Updating vm07:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.conf 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: cephadm 2026-03-07T06:54:41.666460+0000 mgr.vm07.yrfcuj (mgr.14201) 21 : cephadm [INF] Updating vm07:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.conf 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: cephadm 2026-03-07T06:54:41.668627+0000 mgr.vm07.yrfcuj (mgr.14201) 22 : cephadm [INF] Updating vm09:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.conf 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: cephadm 2026-03-07T06:54:41.668627+0000 mgr.vm07.yrfcuj (mgr.14201) 22 : cephadm [INF] Updating vm09:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/config/ceph.conf 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:41.725113+0000 mon.vm07 (mon.0) 266 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:41.725113+0000 mon.vm07 (mon.0) 266 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:41.728468+0000 mon.vm07 (mon.0) 267 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:41.728468+0000 mon.vm07 (mon.0) 267 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:41.732045+0000 mon.vm07 (mon.0) 268 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:41.732045+0000 mon.vm07 (mon.0) 268 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:41.735227+0000 mon.vm07 (mon.0) 269 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:41.735227+0000 mon.vm07 (mon.0) 269 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:41.738801+0000 mon.vm07 (mon.0) 270 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:41.738801+0000 mon.vm07 (mon.0) 270 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: cephadm 2026-03-07T06:54:41.750369+0000 mgr.vm07.yrfcuj (mgr.14201) 23 : cephadm [INF] Reconfiguring mon.vm07 (unknown last config time)... 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: cephadm 2026-03-07T06:54:41.750369+0000 mgr.vm07.yrfcuj (mgr.14201) 23 : cephadm [INF] Reconfiguring mon.vm07 (unknown last config time)... 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:41.750658+0000 mon.vm07 (mon.0) 271 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:41.750658+0000 mon.vm07 (mon.0) 271 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:41.751103+0000 mon.vm07 (mon.0) 272 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:41.751103+0000 mon.vm07 (mon.0) 272 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:41.751444+0000 mon.vm07 (mon.0) 273 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:41.751444+0000 mon.vm07 (mon.0) 273 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: cephadm 2026-03-07T06:54:41.751835+0000 mgr.vm07.yrfcuj (mgr.14201) 24 : cephadm [INF] Reconfiguring daemon mon.vm07 on vm07 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: cephadm 2026-03-07T06:54:41.751835+0000 mgr.vm07.yrfcuj (mgr.14201) 24 : cephadm [INF] Reconfiguring daemon mon.vm07 on vm07 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:42.081935+0000 mon.vm07 (mon.0) 274 : audit [DBG] from='client.? 192.168.123.109:0/529801458' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:42.081935+0000 mon.vm07 (mon.0) 274 : audit [DBG] from='client.? 192.168.123.109:0/529801458' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:42.152767+0000 mon.vm07 (mon.0) 275 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:42.152767+0000 mon.vm07 (mon.0) 275 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:42.156780+0000 mon.vm07 (mon.0) 276 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:42.156780+0000 mon.vm07 (mon.0) 276 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: cephadm 2026-03-07T06:54:42.157472+0000 mgr.vm07.yrfcuj (mgr.14201) 25 : cephadm [INF] Reconfiguring prometheus.vm07 (dependencies changed)... 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: cephadm 2026-03-07T06:54:42.157472+0000 mgr.vm07.yrfcuj (mgr.14201) 25 : cephadm [INF] Reconfiguring prometheus.vm07 (dependencies changed)... 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: cephadm 2026-03-07T06:54:42.323244+0000 mgr.vm07.yrfcuj (mgr.14201) 26 : cephadm [INF] Reconfiguring daemon prometheus.vm07 on vm07 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: cephadm 2026-03-07T06:54:42.323244+0000 mgr.vm07.yrfcuj (mgr.14201) 26 : cephadm [INF] Reconfiguring daemon prometheus.vm07 on vm07 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:42.610058+0000 mon.vm07 (mon.0) 277 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:42.921 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:42 vm07 bash[17031]: audit 2026-03-07T06:54:42.610058+0000 mon.vm07 (mon.0) 277 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:54:44.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:43 vm09 bash[21220]: audit 2026-03-07T06:54:42.943286+0000 mon.vm07 (mon.0) 278 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:44.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:43 vm09 bash[21220]: audit 2026-03-07T06:54:42.943286+0000 mon.vm07 (mon.0) 278 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:44.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:43 vm09 bash[21220]: audit 2026-03-07T06:54:42.948032+0000 mon.vm07 (mon.0) 279 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:44.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:43 vm09 bash[21220]: audit 2026-03-07T06:54:42.948032+0000 mon.vm07 (mon.0) 279 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:44.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:43 vm09 bash[21220]: cephadm 2026-03-07T06:54:42.948857+0000 mgr.vm07.yrfcuj (mgr.14201) 27 : cephadm [INF] Reconfiguring grafana.vm07 (dependencies changed)... 2026-03-07T07:54:44.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:43 vm09 bash[21220]: cephadm 2026-03-07T06:54:42.948857+0000 mgr.vm07.yrfcuj (mgr.14201) 27 : cephadm [INF] Reconfiguring grafana.vm07 (dependencies changed)... 2026-03-07T07:54:44.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:43 vm09 bash[21220]: cephadm 2026-03-07T06:54:42.993584+0000 mgr.vm07.yrfcuj (mgr.14201) 28 : cephadm [INF] Reconfiguring daemon grafana.vm07 on vm07 2026-03-07T07:54:44.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:43 vm09 bash[21220]: cephadm 2026-03-07T06:54:42.993584+0000 mgr.vm07.yrfcuj (mgr.14201) 28 : cephadm [INF] Reconfiguring daemon grafana.vm07 on vm07 2026-03-07T07:54:44.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:43 vm09 bash[21220]: audit 2026-03-07T06:54:43.629892+0000 mon.vm07 (mon.0) 280 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:44.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:43 vm09 bash[21220]: audit 2026-03-07T06:54:43.629892+0000 mon.vm07 (mon.0) 280 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:44.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:43 vm09 bash[21220]: audit 2026-03-07T06:54:43.634442+0000 mon.vm07 (mon.0) 281 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:44.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:43 vm09 bash[21220]: audit 2026-03-07T06:54:43.634442+0000 mon.vm07 (mon.0) 281 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:44.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:43 vm09 bash[21220]: audit 2026-03-07T06:54:43.635403+0000 mon.vm07 (mon.0) 282 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm07", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-07T07:54:44.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:43 vm09 bash[21220]: audit 2026-03-07T06:54:43.635403+0000 mon.vm07 (mon.0) 282 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm07", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-07T07:54:44.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:43 vm09 bash[21220]: audit 2026-03-07T06:54:43.635926+0000 mon.vm07 (mon.0) 283 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:44.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:43 vm09 bash[21220]: audit 2026-03-07T06:54:43.635926+0000 mon.vm07 (mon.0) 283 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:44.291 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:43 vm07 bash[17031]: audit 2026-03-07T06:54:42.943286+0000 mon.vm07 (mon.0) 278 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:44.291 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:43 vm07 bash[17031]: audit 2026-03-07T06:54:42.943286+0000 mon.vm07 (mon.0) 278 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:44.291 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:43 vm07 bash[17031]: audit 2026-03-07T06:54:42.948032+0000 mon.vm07 (mon.0) 279 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:44.291 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:43 vm07 bash[17031]: audit 2026-03-07T06:54:42.948032+0000 mon.vm07 (mon.0) 279 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:44.291 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:43 vm07 bash[17031]: cephadm 2026-03-07T06:54:42.948857+0000 mgr.vm07.yrfcuj (mgr.14201) 27 : cephadm [INF] Reconfiguring grafana.vm07 (dependencies changed)... 2026-03-07T07:54:44.291 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:43 vm07 bash[17031]: cephadm 2026-03-07T06:54:42.948857+0000 mgr.vm07.yrfcuj (mgr.14201) 27 : cephadm [INF] Reconfiguring grafana.vm07 (dependencies changed)... 2026-03-07T07:54:44.291 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:43 vm07 bash[17031]: cephadm 2026-03-07T06:54:42.993584+0000 mgr.vm07.yrfcuj (mgr.14201) 28 : cephadm [INF] Reconfiguring daemon grafana.vm07 on vm07 2026-03-07T07:54:44.291 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:43 vm07 bash[17031]: cephadm 2026-03-07T06:54:42.993584+0000 mgr.vm07.yrfcuj (mgr.14201) 28 : cephadm [INF] Reconfiguring daemon grafana.vm07 on vm07 2026-03-07T07:54:44.291 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:43 vm07 bash[17031]: audit 2026-03-07T06:54:43.629892+0000 mon.vm07 (mon.0) 280 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:44.291 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:43 vm07 bash[17031]: audit 2026-03-07T06:54:43.629892+0000 mon.vm07 (mon.0) 280 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:44.291 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:43 vm07 bash[17031]: audit 2026-03-07T06:54:43.634442+0000 mon.vm07 (mon.0) 281 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:44.291 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:43 vm07 bash[17031]: audit 2026-03-07T06:54:43.634442+0000 mon.vm07 (mon.0) 281 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:44.291 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:43 vm07 bash[17031]: audit 2026-03-07T06:54:43.635403+0000 mon.vm07 (mon.0) 282 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm07", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-07T07:54:44.291 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:43 vm07 bash[17031]: audit 2026-03-07T06:54:43.635403+0000 mon.vm07 (mon.0) 282 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm07", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-07T07:54:44.291 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:43 vm07 bash[17031]: audit 2026-03-07T06:54:43.635926+0000 mon.vm07 (mon.0) 283 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:44.291 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:43 vm07 bash[17031]: audit 2026-03-07T06:54:43.635926+0000 mon.vm07 (mon.0) 283 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:45.047 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: cephadm 2026-03-07T06:54:43.635194+0000 mgr.vm07.yrfcuj (mgr.14201) 29 : cephadm [INF] Reconfiguring crash.vm07 (monmap changed)... 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: cephadm 2026-03-07T06:54:43.635194+0000 mgr.vm07.yrfcuj (mgr.14201) 29 : cephadm [INF] Reconfiguring crash.vm07 (monmap changed)... 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: cephadm 2026-03-07T06:54:43.636356+0000 mgr.vm07.yrfcuj (mgr.14201) 30 : cephadm [INF] Reconfiguring daemon crash.vm07 on vm07 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: cephadm 2026-03-07T06:54:43.636356+0000 mgr.vm07.yrfcuj (mgr.14201) 30 : cephadm [INF] Reconfiguring daemon crash.vm07 on vm07 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.038550+0000 mon.vm07 (mon.0) 284 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.038550+0000 mon.vm07 (mon.0) 284 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.043814+0000 mon.vm07 (mon.0) 285 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.043814+0000 mon.vm07 (mon.0) 285 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: cephadm 2026-03-07T06:54:44.044405+0000 mgr.vm07.yrfcuj (mgr.14201) 31 : cephadm [INF] Reconfiguring mgr.vm07.yrfcuj (unknown last config time)... 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: cephadm 2026-03-07T06:54:44.044405+0000 mgr.vm07.yrfcuj (mgr.14201) 31 : cephadm [INF] Reconfiguring mgr.vm07.yrfcuj (unknown last config time)... 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.044804+0000 mon.vm07 (mon.0) 286 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm07.yrfcuj", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.044804+0000 mon.vm07 (mon.0) 286 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm07.yrfcuj", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.045319+0000 mon.vm07 (mon.0) 287 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.045319+0000 mon.vm07 (mon.0) 287 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.045704+0000 mon.vm07 (mon.0) 288 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.045704+0000 mon.vm07 (mon.0) 288 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: cephadm 2026-03-07T06:54:44.046192+0000 mgr.vm07.yrfcuj (mgr.14201) 32 : cephadm [INF] Reconfiguring daemon mgr.vm07.yrfcuj on vm07 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: cephadm 2026-03-07T06:54:44.046192+0000 mgr.vm07.yrfcuj (mgr.14201) 32 : cephadm [INF] Reconfiguring daemon mgr.vm07.yrfcuj on vm07 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.410358+0000 mon.vm07 (mon.0) 289 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.410358+0000 mon.vm07 (mon.0) 289 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.414595+0000 mon.vm07 (mon.0) 290 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.414595+0000 mon.vm07 (mon.0) 290 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: cephadm 2026-03-07T06:54:44.415261+0000 mgr.vm07.yrfcuj (mgr.14201) 33 : cephadm [INF] Reconfiguring ceph-exporter.vm07 (monmap changed)... 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: cephadm 2026-03-07T06:54:44.415261+0000 mgr.vm07.yrfcuj (mgr.14201) 33 : cephadm [INF] Reconfiguring ceph-exporter.vm07 (monmap changed)... 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.415456+0000 mon.vm07 (mon.0) 291 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm07", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.415456+0000 mon.vm07 (mon.0) 291 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm07", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.415943+0000 mon.vm07 (mon.0) 292 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.415943+0000 mon.vm07 (mon.0) 292 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: cephadm 2026-03-07T06:54:44.416413+0000 mgr.vm07.yrfcuj (mgr.14201) 34 : cephadm [INF] Reconfiguring daemon ceph-exporter.vm07 on vm07 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: cephadm 2026-03-07T06:54:44.416413+0000 mgr.vm07.yrfcuj (mgr.14201) 34 : cephadm [INF] Reconfiguring daemon ceph-exporter.vm07 on vm07 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.799894+0000 mon.vm07 (mon.0) 293 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.799894+0000 mon.vm07 (mon.0) 293 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.803667+0000 mon.vm07 (mon.0) 294 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.048 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:45 vm07 bash[17031]: audit 2026-03-07T06:54:44.803667+0000 mon.vm07 (mon.0) 294 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: cephadm 2026-03-07T06:54:43.635194+0000 mgr.vm07.yrfcuj (mgr.14201) 29 : cephadm [INF] Reconfiguring crash.vm07 (monmap changed)... 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: cephadm 2026-03-07T06:54:43.635194+0000 mgr.vm07.yrfcuj (mgr.14201) 29 : cephadm [INF] Reconfiguring crash.vm07 (monmap changed)... 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: cephadm 2026-03-07T06:54:43.636356+0000 mgr.vm07.yrfcuj (mgr.14201) 30 : cephadm [INF] Reconfiguring daemon crash.vm07 on vm07 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: cephadm 2026-03-07T06:54:43.636356+0000 mgr.vm07.yrfcuj (mgr.14201) 30 : cephadm [INF] Reconfiguring daemon crash.vm07 on vm07 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.038550+0000 mon.vm07 (mon.0) 284 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.038550+0000 mon.vm07 (mon.0) 284 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.043814+0000 mon.vm07 (mon.0) 285 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.043814+0000 mon.vm07 (mon.0) 285 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: cephadm 2026-03-07T06:54:44.044405+0000 mgr.vm07.yrfcuj (mgr.14201) 31 : cephadm [INF] Reconfiguring mgr.vm07.yrfcuj (unknown last config time)... 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: cephadm 2026-03-07T06:54:44.044405+0000 mgr.vm07.yrfcuj (mgr.14201) 31 : cephadm [INF] Reconfiguring mgr.vm07.yrfcuj (unknown last config time)... 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.044804+0000 mon.vm07 (mon.0) 286 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm07.yrfcuj", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.044804+0000 mon.vm07 (mon.0) 286 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm07.yrfcuj", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.045319+0000 mon.vm07 (mon.0) 287 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.045319+0000 mon.vm07 (mon.0) 287 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.045704+0000 mon.vm07 (mon.0) 288 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.045704+0000 mon.vm07 (mon.0) 288 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: cephadm 2026-03-07T06:54:44.046192+0000 mgr.vm07.yrfcuj (mgr.14201) 32 : cephadm [INF] Reconfiguring daemon mgr.vm07.yrfcuj on vm07 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: cephadm 2026-03-07T06:54:44.046192+0000 mgr.vm07.yrfcuj (mgr.14201) 32 : cephadm [INF] Reconfiguring daemon mgr.vm07.yrfcuj on vm07 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.410358+0000 mon.vm07 (mon.0) 289 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.410358+0000 mon.vm07 (mon.0) 289 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.414595+0000 mon.vm07 (mon.0) 290 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.414595+0000 mon.vm07 (mon.0) 290 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: cephadm 2026-03-07T06:54:44.415261+0000 mgr.vm07.yrfcuj (mgr.14201) 33 : cephadm [INF] Reconfiguring ceph-exporter.vm07 (monmap changed)... 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: cephadm 2026-03-07T06:54:44.415261+0000 mgr.vm07.yrfcuj (mgr.14201) 33 : cephadm [INF] Reconfiguring ceph-exporter.vm07 (monmap changed)... 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.415456+0000 mon.vm07 (mon.0) 291 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm07", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.415456+0000 mon.vm07 (mon.0) 291 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm07", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.415943+0000 mon.vm07 (mon.0) 292 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.415943+0000 mon.vm07 (mon.0) 292 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: cephadm 2026-03-07T06:54:44.416413+0000 mgr.vm07.yrfcuj (mgr.14201) 34 : cephadm [INF] Reconfiguring daemon ceph-exporter.vm07 on vm07 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: cephadm 2026-03-07T06:54:44.416413+0000 mgr.vm07.yrfcuj (mgr.14201) 34 : cephadm [INF] Reconfiguring daemon ceph-exporter.vm07 on vm07 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.799894+0000 mon.vm07 (mon.0) 293 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.799894+0000 mon.vm07 (mon.0) 293 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.803667+0000 mon.vm07 (mon.0) 294 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:45 vm09 bash[21220]: audit 2026-03-07T06:54:44.803667+0000 mon.vm07 (mon.0) 294 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: cephadm 2026-03-07T06:54:44.804386+0000 mgr.vm07.yrfcuj (mgr.14201) 35 : cephadm [INF] Reconfiguring alertmanager.vm07 (dependencies changed)... 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: cephadm 2026-03-07T06:54:44.804386+0000 mgr.vm07.yrfcuj (mgr.14201) 35 : cephadm [INF] Reconfiguring alertmanager.vm07 (dependencies changed)... 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: cephadm 2026-03-07T06:54:44.808170+0000 mgr.vm07.yrfcuj (mgr.14201) 36 : cephadm [INF] Reconfiguring daemon alertmanager.vm07 on vm07 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: cephadm 2026-03-07T06:54:44.808170+0000 mgr.vm07.yrfcuj (mgr.14201) 36 : cephadm [INF] Reconfiguring daemon alertmanager.vm07 on vm07 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: cluster 2026-03-07T06:54:45.077902+0000 mgr.vm07.yrfcuj (mgr.14201) 37 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: cluster 2026-03-07T06:54:45.077902+0000 mgr.vm07.yrfcuj (mgr.14201) 37 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.422083+0000 mon.vm07 (mon.0) 295 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.422083+0000 mon.vm07 (mon.0) 295 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.427126+0000 mon.vm07 (mon.0) 296 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.427126+0000 mon.vm07 (mon.0) 296 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: cephadm 2026-03-07T06:54:45.427801+0000 mgr.vm07.yrfcuj (mgr.14201) 38 : cephadm [INF] Reconfiguring ceph-exporter.vm09 (monmap changed)... 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: cephadm 2026-03-07T06:54:45.427801+0000 mgr.vm07.yrfcuj (mgr.14201) 38 : cephadm [INF] Reconfiguring ceph-exporter.vm09 (monmap changed)... 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.428029+0000 mon.vm07 (mon.0) 297 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.428029+0000 mon.vm07 (mon.0) 297 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.428565+0000 mon.vm07 (mon.0) 298 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.428565+0000 mon.vm07 (mon.0) 298 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: cephadm 2026-03-07T06:54:45.429140+0000 mgr.vm07.yrfcuj (mgr.14201) 39 : cephadm [INF] Reconfiguring daemon ceph-exporter.vm09 on vm09 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: cephadm 2026-03-07T06:54:45.429140+0000 mgr.vm07.yrfcuj (mgr.14201) 39 : cephadm [INF] Reconfiguring daemon ceph-exporter.vm09 on vm09 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: cluster 2026-03-07T06:54:45.579685+0000 mon.vm07 (mon.0) 299 : cluster [DBG] Standby manager daemon vm09.eqznpw started 2026-03-07T07:54:46.432 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: cluster 2026-03-07T06:54:45.579685+0000 mon.vm07 (mon.0) 299 : cluster [DBG] Standby manager daemon vm09.eqznpw started 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.582061+0000 mon.vm07 (mon.0) 300 : audit [DBG] from='mgr.? 192.168.123.109:0/3658578978' entity='mgr.vm09.eqznpw' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm09.eqznpw/crt"}]: dispatch 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.582061+0000 mon.vm07 (mon.0) 300 : audit [DBG] from='mgr.? 192.168.123.109:0/3658578978' entity='mgr.vm09.eqznpw' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm09.eqznpw/crt"}]: dispatch 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.587607+0000 mon.vm07 (mon.0) 301 : audit [DBG] from='mgr.? 192.168.123.109:0/3658578978' entity='mgr.vm09.eqznpw' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.587607+0000 mon.vm07 (mon.0) 301 : audit [DBG] from='mgr.? 192.168.123.109:0/3658578978' entity='mgr.vm09.eqznpw' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.590503+0000 mon.vm07 (mon.0) 302 : audit [DBG] from='mgr.? 192.168.123.109:0/3658578978' entity='mgr.vm09.eqznpw' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm09.eqznpw/key"}]: dispatch 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.590503+0000 mon.vm07 (mon.0) 302 : audit [DBG] from='mgr.? 192.168.123.109:0/3658578978' entity='mgr.vm09.eqznpw' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm09.eqznpw/key"}]: dispatch 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.590914+0000 mon.vm07 (mon.0) 303 : audit [DBG] from='mgr.? 192.168.123.109:0/3658578978' entity='mgr.vm09.eqznpw' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.590914+0000 mon.vm07 (mon.0) 303 : audit [DBG] from='mgr.? 192.168.123.109:0/3658578978' entity='mgr.vm09.eqznpw' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.865798+0000 mon.vm07 (mon.0) 304 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.865798+0000 mon.vm07 (mon.0) 304 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.870079+0000 mon.vm07 (mon.0) 305 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.870079+0000 mon.vm07 (mon.0) 305 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.870825+0000 mon.vm07 (mon.0) 306 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.870825+0000 mon.vm07 (mon.0) 306 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.871393+0000 mon.vm07 (mon.0) 307 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.871393+0000 mon.vm07 (mon.0) 307 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.871819+0000 mon.vm07 (mon.0) 308 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:45.871819+0000 mon.vm07 (mon.0) 308 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:46.225025+0000 mon.vm07 (mon.0) 309 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:46.225025+0000 mon.vm07 (mon.0) 309 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:46.228411+0000 mon.vm07 (mon.0) 310 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:46.228411+0000 mon.vm07 (mon.0) 310 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:46.229137+0000 mon.vm07 (mon.0) 311 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:46.229137+0000 mon.vm07 (mon.0) 311 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:46.229680+0000 mon.vm07 (mon.0) 312 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:46.433 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:46 vm09 bash[21220]: audit 2026-03-07T06:54:46.229680+0000 mon.vm07 (mon.0) 312 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:46.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: cephadm 2026-03-07T06:54:44.804386+0000 mgr.vm07.yrfcuj (mgr.14201) 35 : cephadm [INF] Reconfiguring alertmanager.vm07 (dependencies changed)... 2026-03-07T07:54:46.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: cephadm 2026-03-07T06:54:44.804386+0000 mgr.vm07.yrfcuj (mgr.14201) 35 : cephadm [INF] Reconfiguring alertmanager.vm07 (dependencies changed)... 2026-03-07T07:54:46.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: cephadm 2026-03-07T06:54:44.808170+0000 mgr.vm07.yrfcuj (mgr.14201) 36 : cephadm [INF] Reconfiguring daemon alertmanager.vm07 on vm07 2026-03-07T07:54:46.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: cephadm 2026-03-07T06:54:44.808170+0000 mgr.vm07.yrfcuj (mgr.14201) 36 : cephadm [INF] Reconfiguring daemon alertmanager.vm07 on vm07 2026-03-07T07:54:46.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: cluster 2026-03-07T06:54:45.077902+0000 mgr.vm07.yrfcuj (mgr.14201) 37 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:46.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: cluster 2026-03-07T06:54:45.077902+0000 mgr.vm07.yrfcuj (mgr.14201) 37 : cluster [DBG] pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:46.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.422083+0000 mon.vm07 (mon.0) 295 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.422083+0000 mon.vm07 (mon.0) 295 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.427126+0000 mon.vm07 (mon.0) 296 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.427126+0000 mon.vm07 (mon.0) 296 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: cephadm 2026-03-07T06:54:45.427801+0000 mgr.vm07.yrfcuj (mgr.14201) 38 : cephadm [INF] Reconfiguring ceph-exporter.vm09 (monmap changed)... 2026-03-07T07:54:46.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: cephadm 2026-03-07T06:54:45.427801+0000 mgr.vm07.yrfcuj (mgr.14201) 38 : cephadm [INF] Reconfiguring ceph-exporter.vm09 (monmap changed)... 2026-03-07T07:54:46.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.428029+0000 mon.vm07 (mon.0) 297 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-07T07:54:46.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.428029+0000 mon.vm07 (mon.0) 297 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-07T07:54:46.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.428565+0000 mon.vm07 (mon.0) 298 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.428565+0000 mon.vm07 (mon.0) 298 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: cephadm 2026-03-07T06:54:45.429140+0000 mgr.vm07.yrfcuj (mgr.14201) 39 : cephadm [INF] Reconfiguring daemon ceph-exporter.vm09 on vm09 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: cephadm 2026-03-07T06:54:45.429140+0000 mgr.vm07.yrfcuj (mgr.14201) 39 : cephadm [INF] Reconfiguring daemon ceph-exporter.vm09 on vm09 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: cluster 2026-03-07T06:54:45.579685+0000 mon.vm07 (mon.0) 299 : cluster [DBG] Standby manager daemon vm09.eqznpw started 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: cluster 2026-03-07T06:54:45.579685+0000 mon.vm07 (mon.0) 299 : cluster [DBG] Standby manager daemon vm09.eqznpw started 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.582061+0000 mon.vm07 (mon.0) 300 : audit [DBG] from='mgr.? 192.168.123.109:0/3658578978' entity='mgr.vm09.eqznpw' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm09.eqznpw/crt"}]: dispatch 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.582061+0000 mon.vm07 (mon.0) 300 : audit [DBG] from='mgr.? 192.168.123.109:0/3658578978' entity='mgr.vm09.eqznpw' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm09.eqznpw/crt"}]: dispatch 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.587607+0000 mon.vm07 (mon.0) 301 : audit [DBG] from='mgr.? 192.168.123.109:0/3658578978' entity='mgr.vm09.eqznpw' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.587607+0000 mon.vm07 (mon.0) 301 : audit [DBG] from='mgr.? 192.168.123.109:0/3658578978' entity='mgr.vm09.eqznpw' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.590503+0000 mon.vm07 (mon.0) 302 : audit [DBG] from='mgr.? 192.168.123.109:0/3658578978' entity='mgr.vm09.eqznpw' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm09.eqznpw/key"}]: dispatch 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.590503+0000 mon.vm07 (mon.0) 302 : audit [DBG] from='mgr.? 192.168.123.109:0/3658578978' entity='mgr.vm09.eqznpw' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm09.eqznpw/key"}]: dispatch 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.590914+0000 mon.vm07 (mon.0) 303 : audit [DBG] from='mgr.? 192.168.123.109:0/3658578978' entity='mgr.vm09.eqznpw' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.590914+0000 mon.vm07 (mon.0) 303 : audit [DBG] from='mgr.? 192.168.123.109:0/3658578978' entity='mgr.vm09.eqznpw' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.865798+0000 mon.vm07 (mon.0) 304 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.865798+0000 mon.vm07 (mon.0) 304 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.870079+0000 mon.vm07 (mon.0) 305 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.870079+0000 mon.vm07 (mon.0) 305 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.870825+0000 mon.vm07 (mon.0) 306 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.870825+0000 mon.vm07 (mon.0) 306 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.871393+0000 mon.vm07 (mon.0) 307 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.871393+0000 mon.vm07 (mon.0) 307 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.871819+0000 mon.vm07 (mon.0) 308 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:45.871819+0000 mon.vm07 (mon.0) 308 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:46.225025+0000 mon.vm07 (mon.0) 309 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:46.225025+0000 mon.vm07 (mon.0) 309 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:46.228411+0000 mon.vm07 (mon.0) 310 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:46.228411+0000 mon.vm07 (mon.0) 310 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:46.229137+0000 mon.vm07 (mon.0) 311 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:46.229137+0000 mon.vm07 (mon.0) 311 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:46.229680+0000 mon.vm07 (mon.0) 312 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:46.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:46 vm07 bash[17031]: audit 2026-03-07T06:54:46.229680+0000 mon.vm07 (mon.0) 312 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: cephadm 2026-03-07T06:54:45.870661+0000 mgr.vm07.yrfcuj (mgr.14201) 40 : cephadm [INF] Reconfiguring mon.vm09 (monmap changed)... 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: cephadm 2026-03-07T06:54:45.870661+0000 mgr.vm07.yrfcuj (mgr.14201) 40 : cephadm [INF] Reconfiguring mon.vm09 (monmap changed)... 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: cephadm 2026-03-07T06:54:45.872346+0000 mgr.vm07.yrfcuj (mgr.14201) 41 : cephadm [INF] Reconfiguring daemon mon.vm09 on vm09 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: cephadm 2026-03-07T06:54:45.872346+0000 mgr.vm07.yrfcuj (mgr.14201) 41 : cephadm [INF] Reconfiguring daemon mon.vm09 on vm09 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: cephadm 2026-03-07T06:54:46.228949+0000 mgr.vm07.yrfcuj (mgr.14201) 42 : cephadm [INF] Reconfiguring crash.vm09 (monmap changed)... 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: cephadm 2026-03-07T06:54:46.228949+0000 mgr.vm07.yrfcuj (mgr.14201) 42 : cephadm [INF] Reconfiguring crash.vm09 (monmap changed)... 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: cephadm 2026-03-07T06:54:46.230095+0000 mgr.vm07.yrfcuj (mgr.14201) 43 : cephadm [INF] Reconfiguring daemon crash.vm09 on vm09 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: cephadm 2026-03-07T06:54:46.230095+0000 mgr.vm07.yrfcuj (mgr.14201) 43 : cephadm [INF] Reconfiguring daemon crash.vm09 on vm09 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: cluster 2026-03-07T06:54:46.452822+0000 mon.vm07 (mon.0) 313 : cluster [DBG] mgrmap e18: vm07.yrfcuj(active, since 21s), standbys: vm09.eqznpw 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: cluster 2026-03-07T06:54:46.452822+0000 mon.vm07 (mon.0) 313 : cluster [DBG] mgrmap e18: vm07.yrfcuj(active, since 21s), standbys: vm09.eqznpw 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.452926+0000 mon.vm07 (mon.0) 314 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr metadata", "who": "vm09.eqznpw", "id": "vm09.eqznpw"}]: dispatch 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.452926+0000 mon.vm07 (mon.0) 314 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr metadata", "who": "vm09.eqznpw", "id": "vm09.eqznpw"}]: dispatch 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.582550+0000 mon.vm07 (mon.0) 315 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.582550+0000 mon.vm07 (mon.0) 315 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.586514+0000 mon.vm07 (mon.0) 316 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.586514+0000 mon.vm07 (mon.0) 316 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.587287+0000 mon.vm07 (mon.0) 317 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm09.eqznpw", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.587287+0000 mon.vm07 (mon.0) 317 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm09.eqznpw", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.587796+0000 mon.vm07 (mon.0) 318 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T07:54:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.587796+0000 mon.vm07 (mon.0) 318 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.588173+0000 mon.vm07 (mon.0) 319 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.588173+0000 mon.vm07 (mon.0) 319 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.931881+0000 mon.vm07 (mon.0) 320 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.931881+0000 mon.vm07 (mon.0) 320 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.935868+0000 mon.vm07 (mon.0) 321 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.935868+0000 mon.vm07 (mon.0) 321 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.938596+0000 mon.vm07 (mon.0) 322 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.938596+0000 mon.vm07 (mon.0) 322 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.939661+0000 mon.vm07 (mon.0) 323 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm07.local:9095"}]: dispatch 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.939661+0000 mon.vm07 (mon.0) 323 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm07.local:9095"}]: dispatch 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.942644+0000 mon.vm07 (mon.0) 324 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.942644+0000 mon.vm07 (mon.0) 324 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.952247+0000 mon.vm07 (mon.0) 325 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.952247+0000 mon.vm07 (mon.0) 325 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.952889+0000 mon.vm07 (mon.0) 326 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm07.local:3000"}]: dispatch 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.952889+0000 mon.vm07 (mon.0) 326 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm07.local:3000"}]: dispatch 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.956285+0000 mon.vm07 (mon.0) 327 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.956285+0000 mon.vm07 (mon.0) 327 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.962446+0000 mon.vm07 (mon.0) 328 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.962446+0000 mon.vm07 (mon.0) 328 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.963084+0000 mon.vm07 (mon.0) 329 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm07.local:9093"}]: dispatch 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.963084+0000 mon.vm07 (mon.0) 329 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm07.local:9093"}]: dispatch 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.966131+0000 mon.vm07 (mon.0) 330 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.966131+0000 mon.vm07 (mon.0) 330 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.996892+0000 mon.vm07 (mon.0) 331 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:54:47.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:47 vm07 bash[17031]: audit 2026-03-07T06:54:46.996892+0000 mon.vm07 (mon.0) 331 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: cephadm 2026-03-07T06:54:45.870661+0000 mgr.vm07.yrfcuj (mgr.14201) 40 : cephadm [INF] Reconfiguring mon.vm09 (monmap changed)... 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: cephadm 2026-03-07T06:54:45.870661+0000 mgr.vm07.yrfcuj (mgr.14201) 40 : cephadm [INF] Reconfiguring mon.vm09 (monmap changed)... 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: cephadm 2026-03-07T06:54:45.872346+0000 mgr.vm07.yrfcuj (mgr.14201) 41 : cephadm [INF] Reconfiguring daemon mon.vm09 on vm09 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: cephadm 2026-03-07T06:54:45.872346+0000 mgr.vm07.yrfcuj (mgr.14201) 41 : cephadm [INF] Reconfiguring daemon mon.vm09 on vm09 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: cephadm 2026-03-07T06:54:46.228949+0000 mgr.vm07.yrfcuj (mgr.14201) 42 : cephadm [INF] Reconfiguring crash.vm09 (monmap changed)... 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: cephadm 2026-03-07T06:54:46.228949+0000 mgr.vm07.yrfcuj (mgr.14201) 42 : cephadm [INF] Reconfiguring crash.vm09 (monmap changed)... 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: cephadm 2026-03-07T06:54:46.230095+0000 mgr.vm07.yrfcuj (mgr.14201) 43 : cephadm [INF] Reconfiguring daemon crash.vm09 on vm09 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: cephadm 2026-03-07T06:54:46.230095+0000 mgr.vm07.yrfcuj (mgr.14201) 43 : cephadm [INF] Reconfiguring daemon crash.vm09 on vm09 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: cluster 2026-03-07T06:54:46.452822+0000 mon.vm07 (mon.0) 313 : cluster [DBG] mgrmap e18: vm07.yrfcuj(active, since 21s), standbys: vm09.eqznpw 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: cluster 2026-03-07T06:54:46.452822+0000 mon.vm07 (mon.0) 313 : cluster [DBG] mgrmap e18: vm07.yrfcuj(active, since 21s), standbys: vm09.eqznpw 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.452926+0000 mon.vm07 (mon.0) 314 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr metadata", "who": "vm09.eqznpw", "id": "vm09.eqznpw"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.452926+0000 mon.vm07 (mon.0) 314 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr metadata", "who": "vm09.eqznpw", "id": "vm09.eqznpw"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.582550+0000 mon.vm07 (mon.0) 315 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.582550+0000 mon.vm07 (mon.0) 315 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.586514+0000 mon.vm07 (mon.0) 316 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.586514+0000 mon.vm07 (mon.0) 316 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.587287+0000 mon.vm07 (mon.0) 317 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm09.eqznpw", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.587287+0000 mon.vm07 (mon.0) 317 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm09.eqznpw", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.587796+0000 mon.vm07 (mon.0) 318 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.587796+0000 mon.vm07 (mon.0) 318 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.588173+0000 mon.vm07 (mon.0) 319 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.588173+0000 mon.vm07 (mon.0) 319 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.931881+0000 mon.vm07 (mon.0) 320 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.931881+0000 mon.vm07 (mon.0) 320 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.935868+0000 mon.vm07 (mon.0) 321 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.935868+0000 mon.vm07 (mon.0) 321 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.938596+0000 mon.vm07 (mon.0) 322 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.938596+0000 mon.vm07 (mon.0) 322 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.939661+0000 mon.vm07 (mon.0) 323 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm07.local:9095"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.939661+0000 mon.vm07 (mon.0) 323 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm07.local:9095"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.942644+0000 mon.vm07 (mon.0) 324 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.942644+0000 mon.vm07 (mon.0) 324 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.952247+0000 mon.vm07 (mon.0) 325 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.952247+0000 mon.vm07 (mon.0) 325 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.952889+0000 mon.vm07 (mon.0) 326 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm07.local:3000"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.952889+0000 mon.vm07 (mon.0) 326 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm07.local:3000"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.956285+0000 mon.vm07 (mon.0) 327 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.956285+0000 mon.vm07 (mon.0) 327 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.962446+0000 mon.vm07 (mon.0) 328 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.962446+0000 mon.vm07 (mon.0) 328 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.963084+0000 mon.vm07 (mon.0) 329 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm07.local:9093"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.963084+0000 mon.vm07 (mon.0) 329 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm07.local:9093"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.966131+0000 mon.vm07 (mon.0) 330 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.966131+0000 mon.vm07 (mon.0) 330 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.996892+0000 mon.vm07 (mon.0) 331 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:54:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:47 vm09 bash[21220]: audit 2026-03-07T06:54:46.996892+0000 mon.vm07 (mon.0) 331 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:54:48.006 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:54:48.364 INFO:teuthology.orchestra.run.vm07.stdout:# minimal ceph.conf for 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:54:48.364 INFO:teuthology.orchestra.run.vm07.stdout:[global] 2026-03-07T07:54:48.364 INFO:teuthology.orchestra.run.vm07.stdout: fsid = 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T07:54:48.365 INFO:teuthology.orchestra.run.vm07.stdout: mon_host = [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] 2026-03-07T07:54:48.419 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-07T07:54:48.419 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-07T07:54:48.419 DEBUG:teuthology.orchestra.run.vm07:> sudo dd of=/etc/ceph/ceph.conf 2026-03-07T07:54:48.425 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-07T07:54:48.425 DEBUG:teuthology.orchestra.run.vm07:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-07T07:54:48.474 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-07T07:54:48.474 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/ceph.conf 2026-03-07T07:54:48.482 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-07T07:54:48.482 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-07T07:54:48.530 INFO:tasks.cephadm:Deploying OSDs... 2026-03-07T07:54:48.530 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-07T07:54:48.530 DEBUG:teuthology.orchestra.run.vm07:> dd if=/scratch_devs of=/dev/stdout 2026-03-07T07:54:48.533 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-07T07:54:48.533 DEBUG:teuthology.orchestra.run.vm07:> ls /dev/[sv]d? 2026-03-07T07:54:48.578 INFO:teuthology.orchestra.run.vm07.stdout:/dev/vda 2026-03-07T07:54:48.579 INFO:teuthology.orchestra.run.vm07.stdout:/dev/vdb 2026-03-07T07:54:48.579 INFO:teuthology.orchestra.run.vm07.stdout:/dev/vdc 2026-03-07T07:54:48.579 INFO:teuthology.orchestra.run.vm07.stdout:/dev/vdd 2026-03-07T07:54:48.579 INFO:teuthology.orchestra.run.vm07.stdout:/dev/vde 2026-03-07T07:54:48.579 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-07T07:54:48.579 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-07T07:54:48.579 DEBUG:teuthology.orchestra.run.vm07:> stat /dev/vdb 2026-03-07T07:54:48.622 INFO:teuthology.orchestra.run.vm07.stdout: File: /dev/vdb 2026-03-07T07:54:48.622 INFO:teuthology.orchestra.run.vm07.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-07T07:54:48.622 INFO:teuthology.orchestra.run.vm07.stdout:Device: 5h/5d Inode: 24 Links: 1 Device type: fe,10 2026-03-07T07:54:48.622 INFO:teuthology.orchestra.run.vm07.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-07T07:54:48.622 INFO:teuthology.orchestra.run.vm07.stdout:Access: 2026-03-07 07:49:09.639481350 +0100 2026-03-07T07:54:48.622 INFO:teuthology.orchestra.run.vm07.stdout:Modify: 2026-03-07 07:49:08.695481350 +0100 2026-03-07T07:54:48.622 INFO:teuthology.orchestra.run.vm07.stdout:Change: 2026-03-07 07:49:08.695481350 +0100 2026-03-07T07:54:48.623 INFO:teuthology.orchestra.run.vm07.stdout: Birth: - 2026-03-07T07:54:48.623 DEBUG:teuthology.orchestra.run.vm07:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: cephadm 2026-03-07T06:54:46.587000+0000 mgr.vm07.yrfcuj (mgr.14201) 44 : cephadm [INF] Reconfiguring mgr.vm09.eqznpw (monmap changed)... 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: cephadm 2026-03-07T06:54:46.587000+0000 mgr.vm07.yrfcuj (mgr.14201) 44 : cephadm [INF] Reconfiguring mgr.vm09.eqznpw (monmap changed)... 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: cephadm 2026-03-07T06:54:46.588629+0000 mgr.vm07.yrfcuj (mgr.14201) 45 : cephadm [INF] Reconfiguring daemon mgr.vm09.eqznpw on vm09 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: cephadm 2026-03-07T06:54:46.588629+0000 mgr.vm07.yrfcuj (mgr.14201) 45 : cephadm [INF] Reconfiguring daemon mgr.vm09.eqznpw on vm09 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: audit 2026-03-07T06:54:46.939028+0000 mgr.vm07.yrfcuj (mgr.14201) 46 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: audit 2026-03-07T06:54:46.939028+0000 mgr.vm07.yrfcuj (mgr.14201) 46 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: audit 2026-03-07T06:54:46.939885+0000 mgr.vm07.yrfcuj (mgr.14201) 47 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm07.local:9095"}]: dispatch 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: audit 2026-03-07T06:54:46.939885+0000 mgr.vm07.yrfcuj (mgr.14201) 47 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm07.local:9095"}]: dispatch 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: audit 2026-03-07T06:54:46.952457+0000 mgr.vm07.yrfcuj (mgr.14201) 48 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: audit 2026-03-07T06:54:46.952457+0000 mgr.vm07.yrfcuj (mgr.14201) 48 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: audit 2026-03-07T06:54:46.953048+0000 mgr.vm07.yrfcuj (mgr.14201) 49 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm07.local:3000"}]: dispatch 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: audit 2026-03-07T06:54:46.953048+0000 mgr.vm07.yrfcuj (mgr.14201) 49 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm07.local:3000"}]: dispatch 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: audit 2026-03-07T06:54:46.962646+0000 mgr.vm07.yrfcuj (mgr.14201) 50 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: audit 2026-03-07T06:54:46.962646+0000 mgr.vm07.yrfcuj (mgr.14201) 50 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: audit 2026-03-07T06:54:46.963228+0000 mgr.vm07.yrfcuj (mgr.14201) 51 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm07.local:9093"}]: dispatch 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: audit 2026-03-07T06:54:46.963228+0000 mgr.vm07.yrfcuj (mgr.14201) 51 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm07.local:9093"}]: dispatch 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: cluster 2026-03-07T06:54:47.078979+0000 mgr.vm07.yrfcuj (mgr.14201) 52 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: cluster 2026-03-07T06:54:47.078979+0000 mgr.vm07.yrfcuj (mgr.14201) 52 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: audit 2026-03-07T06:54:48.358990+0000 mon.vm07 (mon.0) 332 : audit [DBG] from='client.? 192.168.123.107:0/2200307988' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:48 vm07 bash[17031]: audit 2026-03-07T06:54:48.358990+0000 mon.vm07 (mon.0) 332 : audit [DBG] from='client.? 192.168.123.107:0/2200307988' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:48.664 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records in 2026-03-07T07:54:48.664 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records out 2026-03-07T07:54:48.664 INFO:teuthology.orchestra.run.vm07.stderr:512 bytes copied, 0.00016022 s, 3.2 MB/s 2026-03-07T07:54:48.665 DEBUG:teuthology.orchestra.run.vm07:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-07T07:54:48.711 DEBUG:teuthology.orchestra.run.vm07:> stat /dev/vdc 2026-03-07T07:54:48.758 INFO:teuthology.orchestra.run.vm07.stdout: File: /dev/vdc 2026-03-07T07:54:48.758 INFO:teuthology.orchestra.run.vm07.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-07T07:54:48.758 INFO:teuthology.orchestra.run.vm07.stdout:Device: 5h/5d Inode: 25 Links: 1 Device type: fe,20 2026-03-07T07:54:48.758 INFO:teuthology.orchestra.run.vm07.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-07T07:54:48.758 INFO:teuthology.orchestra.run.vm07.stdout:Access: 2026-03-07 07:49:09.651481350 +0100 2026-03-07T07:54:48.758 INFO:teuthology.orchestra.run.vm07.stdout:Modify: 2026-03-07 07:49:08.691481350 +0100 2026-03-07T07:54:48.759 INFO:teuthology.orchestra.run.vm07.stdout:Change: 2026-03-07 07:49:08.691481350 +0100 2026-03-07T07:54:48.759 INFO:teuthology.orchestra.run.vm07.stdout: Birth: - 2026-03-07T07:54:48.759 DEBUG:teuthology.orchestra.run.vm07:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-07T07:54:48.806 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records in 2026-03-07T07:54:48.806 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records out 2026-03-07T07:54:48.806 INFO:teuthology.orchestra.run.vm07.stderr:512 bytes copied, 0.000107661 s, 4.8 MB/s 2026-03-07T07:54:48.807 DEBUG:teuthology.orchestra.run.vm07:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-07T07:54:48.852 DEBUG:teuthology.orchestra.run.vm07:> stat /dev/vdd 2026-03-07T07:54:48.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: cephadm 2026-03-07T06:54:46.587000+0000 mgr.vm07.yrfcuj (mgr.14201) 44 : cephadm [INF] Reconfiguring mgr.vm09.eqznpw (monmap changed)... 2026-03-07T07:54:48.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: cephadm 2026-03-07T06:54:46.587000+0000 mgr.vm07.yrfcuj (mgr.14201) 44 : cephadm [INF] Reconfiguring mgr.vm09.eqznpw (monmap changed)... 2026-03-07T07:54:48.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: cephadm 2026-03-07T06:54:46.588629+0000 mgr.vm07.yrfcuj (mgr.14201) 45 : cephadm [INF] Reconfiguring daemon mgr.vm09.eqznpw on vm09 2026-03-07T07:54:48.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: cephadm 2026-03-07T06:54:46.588629+0000 mgr.vm07.yrfcuj (mgr.14201) 45 : cephadm [INF] Reconfiguring daemon mgr.vm09.eqznpw on vm09 2026-03-07T07:54:48.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: audit 2026-03-07T06:54:46.939028+0000 mgr.vm07.yrfcuj (mgr.14201) 46 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-07T07:54:48.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: audit 2026-03-07T06:54:46.939028+0000 mgr.vm07.yrfcuj (mgr.14201) 46 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-07T07:54:48.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: audit 2026-03-07T06:54:46.939885+0000 mgr.vm07.yrfcuj (mgr.14201) 47 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm07.local:9095"}]: dispatch 2026-03-07T07:54:48.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: audit 2026-03-07T06:54:46.939885+0000 mgr.vm07.yrfcuj (mgr.14201) 47 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm07.local:9095"}]: dispatch 2026-03-07T07:54:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: audit 2026-03-07T06:54:46.952457+0000 mgr.vm07.yrfcuj (mgr.14201) 48 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-07T07:54:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: audit 2026-03-07T06:54:46.952457+0000 mgr.vm07.yrfcuj (mgr.14201) 48 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-07T07:54:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: audit 2026-03-07T06:54:46.953048+0000 mgr.vm07.yrfcuj (mgr.14201) 49 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm07.local:3000"}]: dispatch 2026-03-07T07:54:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: audit 2026-03-07T06:54:46.953048+0000 mgr.vm07.yrfcuj (mgr.14201) 49 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm07.local:3000"}]: dispatch 2026-03-07T07:54:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: audit 2026-03-07T06:54:46.962646+0000 mgr.vm07.yrfcuj (mgr.14201) 50 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-07T07:54:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: audit 2026-03-07T06:54:46.962646+0000 mgr.vm07.yrfcuj (mgr.14201) 50 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-07T07:54:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: audit 2026-03-07T06:54:46.963228+0000 mgr.vm07.yrfcuj (mgr.14201) 51 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm07.local:9093"}]: dispatch 2026-03-07T07:54:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: audit 2026-03-07T06:54:46.963228+0000 mgr.vm07.yrfcuj (mgr.14201) 51 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm07.local:9093"}]: dispatch 2026-03-07T07:54:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: cluster 2026-03-07T06:54:47.078979+0000 mgr.vm07.yrfcuj (mgr.14201) 52 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: cluster 2026-03-07T06:54:47.078979+0000 mgr.vm07.yrfcuj (mgr.14201) 52 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: audit 2026-03-07T06:54:48.358990+0000 mon.vm07 (mon.0) 332 : audit [DBG] from='client.? 192.168.123.107:0/2200307988' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:48 vm09 bash[21220]: audit 2026-03-07T06:54:48.358990+0000 mon.vm07 (mon.0) 332 : audit [DBG] from='client.? 192.168.123.107:0/2200307988' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:48.898 INFO:teuthology.orchestra.run.vm07.stdout: File: /dev/vdd 2026-03-07T07:54:48.898 INFO:teuthology.orchestra.run.vm07.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-07T07:54:48.898 INFO:teuthology.orchestra.run.vm07.stdout:Device: 5h/5d Inode: 26 Links: 1 Device type: fe,30 2026-03-07T07:54:48.898 INFO:teuthology.orchestra.run.vm07.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-07T07:54:48.898 INFO:teuthology.orchestra.run.vm07.stdout:Access: 2026-03-07 07:49:09.639481350 +0100 2026-03-07T07:54:48.898 INFO:teuthology.orchestra.run.vm07.stdout:Modify: 2026-03-07 07:49:08.699481350 +0100 2026-03-07T07:54:48.898 INFO:teuthology.orchestra.run.vm07.stdout:Change: 2026-03-07 07:49:08.699481350 +0100 2026-03-07T07:54:48.898 INFO:teuthology.orchestra.run.vm07.stdout: Birth: - 2026-03-07T07:54:48.898 DEBUG:teuthology.orchestra.run.vm07:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-07T07:54:48.945 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records in 2026-03-07T07:54:48.945 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records out 2026-03-07T07:54:48.945 INFO:teuthology.orchestra.run.vm07.stderr:512 bytes copied, 0.000107391 s, 4.8 MB/s 2026-03-07T07:54:48.946 DEBUG:teuthology.orchestra.run.vm07:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-07T07:54:48.991 DEBUG:teuthology.orchestra.run.vm07:> stat /dev/vde 2026-03-07T07:54:49.034 INFO:teuthology.orchestra.run.vm07.stdout: File: /dev/vde 2026-03-07T07:54:49.034 INFO:teuthology.orchestra.run.vm07.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-07T07:54:49.034 INFO:teuthology.orchestra.run.vm07.stdout:Device: 5h/5d Inode: 27 Links: 1 Device type: fe,40 2026-03-07T07:54:49.034 INFO:teuthology.orchestra.run.vm07.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-07T07:54:49.034 INFO:teuthology.orchestra.run.vm07.stdout:Access: 2026-03-07 07:49:09.647481350 +0100 2026-03-07T07:54:49.034 INFO:teuthology.orchestra.run.vm07.stdout:Modify: 2026-03-07 07:49:08.719481350 +0100 2026-03-07T07:54:49.034 INFO:teuthology.orchestra.run.vm07.stdout:Change: 2026-03-07 07:49:08.719481350 +0100 2026-03-07T07:54:49.034 INFO:teuthology.orchestra.run.vm07.stdout: Birth: - 2026-03-07T07:54:49.034 DEBUG:teuthology.orchestra.run.vm07:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-07T07:54:49.082 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records in 2026-03-07T07:54:49.083 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records out 2026-03-07T07:54:49.083 INFO:teuthology.orchestra.run.vm07.stderr:512 bytes copied, 0.000119343 s, 4.3 MB/s 2026-03-07T07:54:49.083 DEBUG:teuthology.orchestra.run.vm07:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-07T07:54:49.127 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-07T07:54:49.127 DEBUG:teuthology.orchestra.run.vm09:> dd if=/scratch_devs of=/dev/stdout 2026-03-07T07:54:49.130 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-07T07:54:49.130 DEBUG:teuthology.orchestra.run.vm09:> ls /dev/[sv]d? 2026-03-07T07:54:49.173 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vda 2026-03-07T07:54:49.174 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vdb 2026-03-07T07:54:49.174 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vdc 2026-03-07T07:54:49.174 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vdd 2026-03-07T07:54:49.174 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vde 2026-03-07T07:54:49.174 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-07T07:54:49.174 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-07T07:54:49.174 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vdb 2026-03-07T07:54:49.218 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vdb 2026-03-07T07:54:49.218 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-07T07:54:49.218 INFO:teuthology.orchestra.run.vm09.stdout:Device: 5h/5d Inode: 24 Links: 1 Device type: fe,10 2026-03-07T07:54:49.218 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-07T07:54:49.218 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-07 07:49:34.774835986 +0100 2026-03-07T07:54:49.218 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-07 07:49:33.698835986 +0100 2026-03-07T07:54:49.218 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-07 07:49:33.698835986 +0100 2026-03-07T07:54:49.218 INFO:teuthology.orchestra.run.vm09.stdout: Birth: - 2026-03-07T07:54:49.218 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-07T07:54:49.265 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-07T07:54:49.266 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-07T07:54:49.266 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000196217 s, 2.6 MB/s 2026-03-07T07:54:49.266 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-07T07:54:49.310 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vdc 2026-03-07T07:54:49.353 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vdc 2026-03-07T07:54:49.353 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-07T07:54:49.353 INFO:teuthology.orchestra.run.vm09.stdout:Device: 5h/5d Inode: 25 Links: 1 Device type: fe,20 2026-03-07T07:54:49.353 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-07T07:54:49.353 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-07 07:49:34.782835986 +0100 2026-03-07T07:54:49.354 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-07 07:49:33.694835986 +0100 2026-03-07T07:54:49.354 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-07 07:49:33.694835986 +0100 2026-03-07T07:54:49.354 INFO:teuthology.orchestra.run.vm09.stdout: Birth: - 2026-03-07T07:54:49.354 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-07T07:54:49.401 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-07T07:54:49.401 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-07T07:54:49.401 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000179797 s, 2.8 MB/s 2026-03-07T07:54:49.402 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-07T07:54:49.447 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vdd 2026-03-07T07:54:49.494 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vdd 2026-03-07T07:54:49.494 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-07T07:54:49.494 INFO:teuthology.orchestra.run.vm09.stdout:Device: 5h/5d Inode: 26 Links: 1 Device type: fe,30 2026-03-07T07:54:49.494 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-07T07:54:49.494 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-07 07:49:34.774835986 +0100 2026-03-07T07:54:49.494 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-07 07:49:33.698835986 +0100 2026-03-07T07:54:49.494 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-07 07:49:33.698835986 +0100 2026-03-07T07:54:49.494 INFO:teuthology.orchestra.run.vm09.stdout: Birth: - 2026-03-07T07:54:49.494 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-07T07:54:49.542 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-07T07:54:49.542 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-07T07:54:49.542 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000166081 s, 3.1 MB/s 2026-03-07T07:54:49.542 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-07T07:54:49.590 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vde 2026-03-07T07:54:49.633 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vde 2026-03-07T07:54:49.633 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 4096 block special file 2026-03-07T07:54:49.633 INFO:teuthology.orchestra.run.vm09.stdout:Device: 5h/5d Inode: 27 Links: 1 Device type: fe,40 2026-03-07T07:54:49.633 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-07T07:54:49.633 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-07 07:49:34.782835986 +0100 2026-03-07T07:54:49.634 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-07 07:49:33.698835986 +0100 2026-03-07T07:54:49.634 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-07 07:49:33.698835986 +0100 2026-03-07T07:54:49.634 INFO:teuthology.orchestra.run.vm09.stdout: Birth: - 2026-03-07T07:54:49.634 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-07T07:54:49.682 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-07T07:54:49.682 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-07T07:54:49.682 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000162625 s, 3.1 MB/s 2026-03-07T07:54:49.682 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-07T07:54:49.727 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph orch apply osd --all-available-devices 2026-03-07T07:54:50.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:50 vm09 bash[21220]: cluster 2026-03-07T06:54:49.079191+0000 mgr.vm07.yrfcuj (mgr.14201) 53 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:50 vm09 bash[21220]: cluster 2026-03-07T06:54:49.079191+0000 mgr.vm07.yrfcuj (mgr.14201) 53 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:50.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:50 vm07 bash[17031]: cluster 2026-03-07T06:54:49.079191+0000 mgr.vm07.yrfcuj (mgr.14201) 53 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:50.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:50 vm07 bash[17031]: cluster 2026-03-07T06:54:49.079191+0000 mgr.vm07.yrfcuj (mgr.14201) 53 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:52 vm09 bash[21220]: cluster 2026-03-07T06:54:51.079346+0000 mgr.vm07.yrfcuj (mgr.14201) 54 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:52 vm09 bash[21220]: cluster 2026-03-07T06:54:51.079346+0000 mgr.vm07.yrfcuj (mgr.14201) 54 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:52 vm09 bash[21220]: audit 2026-03-07T06:54:51.515257+0000 mon.vm07 (mon.0) 333 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:52 vm09 bash[21220]: audit 2026-03-07T06:54:51.515257+0000 mon.vm07 (mon.0) 333 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:52 vm09 bash[21220]: audit 2026-03-07T06:54:51.520758+0000 mon.vm07 (mon.0) 334 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:52 vm09 bash[21220]: audit 2026-03-07T06:54:51.520758+0000 mon.vm07 (mon.0) 334 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:52 vm09 bash[21220]: audit 2026-03-07T06:54:52.023578+0000 mon.vm07 (mon.0) 335 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:52 vm09 bash[21220]: audit 2026-03-07T06:54:52.023578+0000 mon.vm07 (mon.0) 335 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:52 vm09 bash[21220]: audit 2026-03-07T06:54:52.028275+0000 mon.vm07 (mon.0) 336 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:52 vm09 bash[21220]: audit 2026-03-07T06:54:52.028275+0000 mon.vm07 (mon.0) 336 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:52 vm09 bash[21220]: audit 2026-03-07T06:54:52.029171+0000 mon.vm07 (mon.0) 337 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:52 vm09 bash[21220]: audit 2026-03-07T06:54:52.029171+0000 mon.vm07 (mon.0) 337 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:52 vm09 bash[21220]: audit 2026-03-07T06:54:52.029740+0000 mon.vm07 (mon.0) 338 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:54:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:52 vm09 bash[21220]: audit 2026-03-07T06:54:52.029740+0000 mon.vm07 (mon.0) 338 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:54:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:52 vm09 bash[21220]: audit 2026-03-07T06:54:52.033222+0000 mon.vm07 (mon.0) 339 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:52 vm09 bash[21220]: audit 2026-03-07T06:54:52.033222+0000 mon.vm07 (mon.0) 339 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:52 vm07 bash[17031]: cluster 2026-03-07T06:54:51.079346+0000 mgr.vm07.yrfcuj (mgr.14201) 54 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:52 vm07 bash[17031]: cluster 2026-03-07T06:54:51.079346+0000 mgr.vm07.yrfcuj (mgr.14201) 54 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:52 vm07 bash[17031]: audit 2026-03-07T06:54:51.515257+0000 mon.vm07 (mon.0) 333 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:52 vm07 bash[17031]: audit 2026-03-07T06:54:51.515257+0000 mon.vm07 (mon.0) 333 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:52 vm07 bash[17031]: audit 2026-03-07T06:54:51.520758+0000 mon.vm07 (mon.0) 334 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:52 vm07 bash[17031]: audit 2026-03-07T06:54:51.520758+0000 mon.vm07 (mon.0) 334 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:52 vm07 bash[17031]: audit 2026-03-07T06:54:52.023578+0000 mon.vm07 (mon.0) 335 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:52 vm07 bash[17031]: audit 2026-03-07T06:54:52.023578+0000 mon.vm07 (mon.0) 335 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:52 vm07 bash[17031]: audit 2026-03-07T06:54:52.028275+0000 mon.vm07 (mon.0) 336 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:52 vm07 bash[17031]: audit 2026-03-07T06:54:52.028275+0000 mon.vm07 (mon.0) 336 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:52 vm07 bash[17031]: audit 2026-03-07T06:54:52.029171+0000 mon.vm07 (mon.0) 337 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:52 vm07 bash[17031]: audit 2026-03-07T06:54:52.029171+0000 mon.vm07 (mon.0) 337 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:54:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:52 vm07 bash[17031]: audit 2026-03-07T06:54:52.029740+0000 mon.vm07 (mon.0) 338 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:54:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:52 vm07 bash[17031]: audit 2026-03-07T06:54:52.029740+0000 mon.vm07 (mon.0) 338 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:54:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:52 vm07 bash[17031]: audit 2026-03-07T06:54:52.033222+0000 mon.vm07 (mon.0) 339 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:52 vm07 bash[17031]: audit 2026-03-07T06:54:52.033222+0000 mon.vm07 (mon.0) 339 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:53.766 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm09/config 2026-03-07T07:54:54.127 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled osd.all-available-devices update... 2026-03-07T07:54:54.197 INFO:tasks.cephadm:Waiting for 8 OSDs to come up... 2026-03-07T07:54:54.197 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd stat -f json 2026-03-07T07:54:54.777 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:54 vm09 bash[21220]: cluster 2026-03-07T06:54:53.079489+0000 mgr.vm07.yrfcuj (mgr.14201) 55 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:54.777 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:54 vm09 bash[21220]: cluster 2026-03-07T06:54:53.079489+0000 mgr.vm07.yrfcuj (mgr.14201) 55 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:54.777 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:54 vm09 bash[21220]: audit 2026-03-07T06:54:54.121027+0000 mon.vm07 (mon.0) 340 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:54.777 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:54 vm09 bash[21220]: audit 2026-03-07T06:54:54.121027+0000 mon.vm07 (mon.0) 340 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:54.778 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:54 vm09 bash[21220]: audit 2026-03-07T06:54:54.122162+0000 mon.vm07 (mon.0) 341 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:54:54.778 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:54 vm09 bash[21220]: audit 2026-03-07T06:54:54.122162+0000 mon.vm07 (mon.0) 341 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:54:54.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:54 vm07 bash[17031]: cluster 2026-03-07T06:54:53.079489+0000 mgr.vm07.yrfcuj (mgr.14201) 55 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:54.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:54 vm07 bash[17031]: cluster 2026-03-07T06:54:53.079489+0000 mgr.vm07.yrfcuj (mgr.14201) 55 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:54.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:54 vm07 bash[17031]: audit 2026-03-07T06:54:54.121027+0000 mon.vm07 (mon.0) 340 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:54.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:54 vm07 bash[17031]: audit 2026-03-07T06:54:54.121027+0000 mon.vm07 (mon.0) 340 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:54:54.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:54 vm07 bash[17031]: audit 2026-03-07T06:54:54.122162+0000 mon.vm07 (mon.0) 341 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:54:54.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:54 vm07 bash[17031]: audit 2026-03-07T06:54:54.122162+0000 mon.vm07 (mon.0) 341 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:54:55.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:55 vm09 bash[21220]: audit 2026-03-07T06:54:54.115259+0000 mgr.vm07.yrfcuj (mgr.14201) 56 : audit [DBG] from='client.14232 -' entity='client.admin' cmd=[{"prefix": "orch apply osd", "all_available_devices": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:54:55.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:55 vm09 bash[21220]: audit 2026-03-07T06:54:54.115259+0000 mgr.vm07.yrfcuj (mgr.14201) 56 : audit [DBG] from='client.14232 -' entity='client.admin' cmd=[{"prefix": "orch apply osd", "all_available_devices": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:54:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:55 vm09 bash[21220]: cephadm 2026-03-07T06:54:54.116364+0000 mgr.vm07.yrfcuj (mgr.14201) 57 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-07T07:54:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:55 vm09 bash[21220]: cephadm 2026-03-07T06:54:54.116364+0000 mgr.vm07.yrfcuj (mgr.14201) 57 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-07T07:54:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:55 vm09 bash[21220]: cephadm 2026-03-07T06:54:54.116382+0000 mgr.vm07.yrfcuj (mgr.14201) 58 : cephadm [INF] Marking host: vm09 for OSDSpec preview refresh. 2026-03-07T07:54:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:55 vm09 bash[21220]: cephadm 2026-03-07T06:54:54.116382+0000 mgr.vm07.yrfcuj (mgr.14201) 58 : cephadm [INF] Marking host: vm09 for OSDSpec preview refresh. 2026-03-07T07:54:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:55 vm09 bash[21220]: cephadm 2026-03-07T06:54:54.116544+0000 mgr.vm07.yrfcuj (mgr.14201) 59 : cephadm [INF] Saving service osd.all-available-devices spec with placement * 2026-03-07T07:54:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:55 vm09 bash[21220]: cephadm 2026-03-07T06:54:54.116544+0000 mgr.vm07.yrfcuj (mgr.14201) 59 : cephadm [INF] Saving service osd.all-available-devices spec with placement * 2026-03-07T07:54:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:55 vm09 bash[21220]: audit 2026-03-07T06:54:55.124299+0000 mon.vm07 (mon.0) 342 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:54:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:55 vm09 bash[21220]: audit 2026-03-07T06:54:55.124299+0000 mon.vm07 (mon.0) 342 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:54:55.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:55 vm07 bash[17031]: audit 2026-03-07T06:54:54.115259+0000 mgr.vm07.yrfcuj (mgr.14201) 56 : audit [DBG] from='client.14232 -' entity='client.admin' cmd=[{"prefix": "orch apply osd", "all_available_devices": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:54:55.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:55 vm07 bash[17031]: audit 2026-03-07T06:54:54.115259+0000 mgr.vm07.yrfcuj (mgr.14201) 56 : audit [DBG] from='client.14232 -' entity='client.admin' cmd=[{"prefix": "orch apply osd", "all_available_devices": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:54:55.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:55 vm07 bash[17031]: cephadm 2026-03-07T06:54:54.116364+0000 mgr.vm07.yrfcuj (mgr.14201) 57 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-07T07:54:55.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:55 vm07 bash[17031]: cephadm 2026-03-07T06:54:54.116364+0000 mgr.vm07.yrfcuj (mgr.14201) 57 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-07T07:54:55.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:55 vm07 bash[17031]: cephadm 2026-03-07T06:54:54.116382+0000 mgr.vm07.yrfcuj (mgr.14201) 58 : cephadm [INF] Marking host: vm09 for OSDSpec preview refresh. 2026-03-07T07:54:55.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:55 vm07 bash[17031]: cephadm 2026-03-07T06:54:54.116382+0000 mgr.vm07.yrfcuj (mgr.14201) 58 : cephadm [INF] Marking host: vm09 for OSDSpec preview refresh. 2026-03-07T07:54:55.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:55 vm07 bash[17031]: cephadm 2026-03-07T06:54:54.116544+0000 mgr.vm07.yrfcuj (mgr.14201) 59 : cephadm [INF] Saving service osd.all-available-devices spec with placement * 2026-03-07T07:54:55.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:55 vm07 bash[17031]: cephadm 2026-03-07T06:54:54.116544+0000 mgr.vm07.yrfcuj (mgr.14201) 59 : cephadm [INF] Saving service osd.all-available-devices spec with placement * 2026-03-07T07:54:55.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:55 vm07 bash[17031]: audit 2026-03-07T06:54:55.124299+0000 mon.vm07 (mon.0) 342 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:54:55.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:55 vm07 bash[17031]: audit 2026-03-07T06:54:55.124299+0000 mon.vm07 (mon.0) 342 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:54:56.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:56 vm09 bash[21220]: cluster 2026-03-07T06:54:55.079680+0000 mgr.vm07.yrfcuj (mgr.14201) 60 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:56.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:56 vm09 bash[21220]: cluster 2026-03-07T06:54:55.079680+0000 mgr.vm07.yrfcuj (mgr.14201) 60 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:56.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:56 vm07 bash[17031]: cluster 2026-03-07T06:54:55.079680+0000 mgr.vm07.yrfcuj (mgr.14201) 60 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:56.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:56 vm07 bash[17031]: cluster 2026-03-07T06:54:55.079680+0000 mgr.vm07.yrfcuj (mgr.14201) 60 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:57.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:57 vm09 bash[21220]: cluster 2026-03-07T06:54:57.080159+0000 mgr.vm07.yrfcuj (mgr.14201) 61 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:57.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:54:57 vm09 bash[21220]: cluster 2026-03-07T06:54:57.080159+0000 mgr.vm07.yrfcuj (mgr.14201) 61 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:57.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:57 vm07 bash[17031]: cluster 2026-03-07T06:54:57.080159+0000 mgr.vm07.yrfcuj (mgr.14201) 61 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:57.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:54:57 vm07 bash[17031]: cluster 2026-03-07T06:54:57.080159+0000 mgr.vm07.yrfcuj (mgr.14201) 61 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:54:58.910 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:54:59.380 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:54:59.470 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":5,"num_osds":0,"num_up_osds":0,"osd_up_since":0,"num_in_osds":0,"osd_in_since":0,"num_remapped_pgs":0} 2026-03-07T07:55:00.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: cluster 2026-03-07T06:54:59.080515+0000 mgr.vm07.yrfcuj (mgr.14201) 62 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:00.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: cluster 2026-03-07T06:54:59.080515+0000 mgr.vm07.yrfcuj (mgr.14201) 62 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:00.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.144650+0000 mon.vm07 (mon.0) 343 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.144650+0000 mon.vm07 (mon.0) 343 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.149492+0000 mon.vm07 (mon.0) 344 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.149492+0000 mon.vm07 (mon.0) 344 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.156573+0000 mon.vm07 (mon.0) 345 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.156573+0000 mon.vm07 (mon.0) 345 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.162255+0000 mon.vm07 (mon.0) 346 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.162255+0000 mon.vm07 (mon.0) 346 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.375033+0000 mon.vm07 (mon.0) 347 : audit [DBG] from='client.? 192.168.123.107:0/3127656195' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.375033+0000 mon.vm07 (mon.0) 347 : audit [DBG] from='client.? 192.168.123.107:0/3127656195' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.597688+0000 mon.vm07 (mon.0) 348 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.597688+0000 mon.vm07 (mon.0) 348 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.601173+0000 mon.vm07 (mon.0) 349 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.601173+0000 mon.vm07 (mon.0) 349 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.604338+0000 mon.vm07 (mon.0) 350 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.604338+0000 mon.vm07 (mon.0) 350 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.607676+0000 mon.vm07 (mon.0) 351 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.607676+0000 mon.vm07 (mon.0) 351 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.608271+0000 mon.vm07 (mon.0) 352 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.608271+0000 mon.vm07 (mon.0) 352 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.608705+0000 mon.vm07 (mon.0) 353 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.608705+0000 mon.vm07 (mon.0) 353 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.611527+0000 mon.vm07 (mon.0) 354 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.611527+0000 mon.vm07 (mon.0) 354 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.612685+0000 mon.vm07 (mon.0) 355 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.612685+0000 mon.vm07 (mon.0) 355 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.614505+0000 mon.vm07 (mon.0) 356 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.614505+0000 mon.vm07 (mon.0) 356 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.614906+0000 mon.vm07 (mon.0) 357 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.614906+0000 mon.vm07 (mon.0) 357 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.616354+0000 mon.vm07 (mon.0) 358 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.616354+0000 mon.vm07 (mon.0) 358 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.616756+0000 mon.vm07 (mon.0) 359 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:00.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:00 vm07 bash[17031]: audit 2026-03-07T06:54:59.616756+0000 mon.vm07 (mon.0) 359 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:00.471 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd stat -f json 2026-03-07T07:55:00.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: cluster 2026-03-07T06:54:59.080515+0000 mgr.vm07.yrfcuj (mgr.14201) 62 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:00.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: cluster 2026-03-07T06:54:59.080515+0000 mgr.vm07.yrfcuj (mgr.14201) 62 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:00.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.144650+0000 mon.vm07 (mon.0) 343 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.144650+0000 mon.vm07 (mon.0) 343 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.149492+0000 mon.vm07 (mon.0) 344 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.149492+0000 mon.vm07 (mon.0) 344 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.156573+0000 mon.vm07 (mon.0) 345 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.156573+0000 mon.vm07 (mon.0) 345 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.162255+0000 mon.vm07 (mon.0) 346 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.162255+0000 mon.vm07 (mon.0) 346 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.375033+0000 mon.vm07 (mon.0) 347 : audit [DBG] from='client.? 192.168.123.107:0/3127656195' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.375033+0000 mon.vm07 (mon.0) 347 : audit [DBG] from='client.? 192.168.123.107:0/3127656195' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.597688+0000 mon.vm07 (mon.0) 348 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.597688+0000 mon.vm07 (mon.0) 348 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.601173+0000 mon.vm07 (mon.0) 349 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.601173+0000 mon.vm07 (mon.0) 349 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.604338+0000 mon.vm07 (mon.0) 350 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.604338+0000 mon.vm07 (mon.0) 350 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.607676+0000 mon.vm07 (mon.0) 351 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.607676+0000 mon.vm07 (mon.0) 351 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.608271+0000 mon.vm07 (mon.0) 352 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.608271+0000 mon.vm07 (mon.0) 352 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.608705+0000 mon.vm07 (mon.0) 353 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.608705+0000 mon.vm07 (mon.0) 353 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.611527+0000 mon.vm07 (mon.0) 354 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.611527+0000 mon.vm07 (mon.0) 354 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.612685+0000 mon.vm07 (mon.0) 355 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.612685+0000 mon.vm07 (mon.0) 355 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.614505+0000 mon.vm07 (mon.0) 356 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.614505+0000 mon.vm07 (mon.0) 356 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.614906+0000 mon.vm07 (mon.0) 357 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.614906+0000 mon.vm07 (mon.0) 357 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.616354+0000 mon.vm07 (mon.0) 358 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.616354+0000 mon.vm07 (mon.0) 358 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.616756+0000 mon.vm07 (mon.0) 359 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:00 vm09 bash[21220]: audit 2026-03-07T06:54:59.616756+0000 mon.vm07 (mon.0) 359 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:02.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:02 vm07 bash[17031]: cluster 2026-03-07T06:55:01.081044+0000 mgr.vm07.yrfcuj (mgr.14201) 63 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:02.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:02 vm07 bash[17031]: cluster 2026-03-07T06:55:01.081044+0000 mgr.vm07.yrfcuj (mgr.14201) 63 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:02.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:02 vm09 bash[21220]: cluster 2026-03-07T06:55:01.081044+0000 mgr.vm07.yrfcuj (mgr.14201) 63 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:02.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:02 vm09 bash[21220]: cluster 2026-03-07T06:55:01.081044+0000 mgr.vm07.yrfcuj (mgr.14201) 63 : cluster [DBG] pgmap v11: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:04.396 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:55:04.420 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:04 vm07 bash[17031]: cluster 2026-03-07T06:55:03.081325+0000 mgr.vm07.yrfcuj (mgr.14201) 64 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:04.420 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:04 vm07 bash[17031]: cluster 2026-03-07T06:55:03.081325+0000 mgr.vm07.yrfcuj (mgr.14201) 64 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:04.427 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:04 vm09 bash[21220]: cluster 2026-03-07T06:55:03.081325+0000 mgr.vm07.yrfcuj (mgr.14201) 64 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:04.427 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:04 vm09 bash[21220]: cluster 2026-03-07T06:55:03.081325+0000 mgr.vm07.yrfcuj (mgr.14201) 64 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:04.812 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:55:04.911 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":5,"num_osds":0,"num_up_osds":0,"osd_up_since":0,"num_in_osds":0,"osd_in_since":0,"num_remapped_pgs":0} 2026-03-07T07:55:05.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:05 vm07 bash[17031]: audit 2026-03-07T06:55:04.804954+0000 mon.vm07 (mon.0) 360 : audit [DBG] from='client.? 192.168.123.107:0/4036926492' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:05.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:05 vm07 bash[17031]: audit 2026-03-07T06:55:04.804954+0000 mon.vm07 (mon.0) 360 : audit [DBG] from='client.? 192.168.123.107:0/4036926492' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:05.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:05 vm09 bash[21220]: audit 2026-03-07T06:55:04.804954+0000 mon.vm07 (mon.0) 360 : audit [DBG] from='client.? 192.168.123.107:0/4036926492' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:05.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:05 vm09 bash[21220]: audit 2026-03-07T06:55:04.804954+0000 mon.vm07 (mon.0) 360 : audit [DBG] from='client.? 192.168.123.107:0/4036926492' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:05.912 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd stat -f json 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: cluster 2026-03-07T06:55:05.081549+0000 mgr.vm07.yrfcuj (mgr.14201) 65 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: cluster 2026-03-07T06:55:05.081549+0000 mgr.vm07.yrfcuj (mgr.14201) 65 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: audit 2026-03-07T06:55:05.813315+0000 mon.vm07 (mon.0) 361 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e6429ca8-de99-4811-a21e-2292aedf25ef"}]: dispatch 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: audit 2026-03-07T06:55:05.813315+0000 mon.vm07 (mon.0) 361 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e6429ca8-de99-4811-a21e-2292aedf25ef"}]: dispatch 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: audit 2026-03-07T06:55:05.816288+0000 mon.vm07 (mon.0) 362 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "e6429ca8-de99-4811-a21e-2292aedf25ef"}]': finished 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: audit 2026-03-07T06:55:05.816288+0000 mon.vm07 (mon.0) 362 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "e6429ca8-de99-4811-a21e-2292aedf25ef"}]': finished 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: audit 2026-03-07T06:55:05.817623+0000 mon.vm09 (mon.1) 2 : audit [INF] from='client.? 192.168.123.109:0/1770037392' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e6429ca8-de99-4811-a21e-2292aedf25ef"}]: dispatch 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: audit 2026-03-07T06:55:05.817623+0000 mon.vm09 (mon.1) 2 : audit [INF] from='client.? 192.168.123.109:0/1770037392' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e6429ca8-de99-4811-a21e-2292aedf25ef"}]: dispatch 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: cluster 2026-03-07T06:55:05.818267+0000 mon.vm07 (mon.0) 363 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: cluster 2026-03-07T06:55:05.818267+0000 mon.vm07 (mon.0) 363 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: audit 2026-03-07T06:55:05.818591+0000 mon.vm07 (mon.0) 364 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: audit 2026-03-07T06:55:05.818591+0000 mon.vm07 (mon.0) 364 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: audit 2026-03-07T06:55:05.870737+0000 mon.vm07 (mon.0) 365 : audit [INF] from='client.? 192.168.123.107:0/298761861' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "56cba12f-96c0-445c-be10-c84fa665caa4"}]: dispatch 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: audit 2026-03-07T06:55:05.870737+0000 mon.vm07 (mon.0) 365 : audit [INF] from='client.? 192.168.123.107:0/298761861' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "56cba12f-96c0-445c-be10-c84fa665caa4"}]: dispatch 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: audit 2026-03-07T06:55:05.873463+0000 mon.vm07 (mon.0) 366 : audit [INF] from='client.? 192.168.123.107:0/298761861' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "56cba12f-96c0-445c-be10-c84fa665caa4"}]': finished 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: audit 2026-03-07T06:55:05.873463+0000 mon.vm07 (mon.0) 366 : audit [INF] from='client.? 192.168.123.107:0/298761861' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "56cba12f-96c0-445c-be10-c84fa665caa4"}]': finished 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: cluster 2026-03-07T06:55:05.876088+0000 mon.vm07 (mon.0) 367 : cluster [DBG] osdmap e7: 2 total, 0 up, 2 in 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: cluster 2026-03-07T06:55:05.876088+0000 mon.vm07 (mon.0) 367 : cluster [DBG] osdmap e7: 2 total, 0 up, 2 in 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: audit 2026-03-07T06:55:05.876234+0000 mon.vm07 (mon.0) 368 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: audit 2026-03-07T06:55:05.876234+0000 mon.vm07 (mon.0) 368 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: audit 2026-03-07T06:55:05.876297+0000 mon.vm07 (mon.0) 369 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:06.181 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:06 vm07 bash[17031]: audit 2026-03-07T06:55:05.876297+0000 mon.vm07 (mon.0) 369 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:06.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: cluster 2026-03-07T06:55:05.081549+0000 mgr.vm07.yrfcuj (mgr.14201) 65 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: cluster 2026-03-07T06:55:05.081549+0000 mgr.vm07.yrfcuj (mgr.14201) 65 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: audit 2026-03-07T06:55:05.813315+0000 mon.vm07 (mon.0) 361 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e6429ca8-de99-4811-a21e-2292aedf25ef"}]: dispatch 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: audit 2026-03-07T06:55:05.813315+0000 mon.vm07 (mon.0) 361 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e6429ca8-de99-4811-a21e-2292aedf25ef"}]: dispatch 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: audit 2026-03-07T06:55:05.816288+0000 mon.vm07 (mon.0) 362 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "e6429ca8-de99-4811-a21e-2292aedf25ef"}]': finished 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: audit 2026-03-07T06:55:05.816288+0000 mon.vm07 (mon.0) 362 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "e6429ca8-de99-4811-a21e-2292aedf25ef"}]': finished 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: audit 2026-03-07T06:55:05.817623+0000 mon.vm09 (mon.1) 2 : audit [INF] from='client.? 192.168.123.109:0/1770037392' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e6429ca8-de99-4811-a21e-2292aedf25ef"}]: dispatch 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: audit 2026-03-07T06:55:05.817623+0000 mon.vm09 (mon.1) 2 : audit [INF] from='client.? 192.168.123.109:0/1770037392' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e6429ca8-de99-4811-a21e-2292aedf25ef"}]: dispatch 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: cluster 2026-03-07T06:55:05.818267+0000 mon.vm07 (mon.0) 363 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: cluster 2026-03-07T06:55:05.818267+0000 mon.vm07 (mon.0) 363 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: audit 2026-03-07T06:55:05.818591+0000 mon.vm07 (mon.0) 364 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: audit 2026-03-07T06:55:05.818591+0000 mon.vm07 (mon.0) 364 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: audit 2026-03-07T06:55:05.870737+0000 mon.vm07 (mon.0) 365 : audit [INF] from='client.? 192.168.123.107:0/298761861' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "56cba12f-96c0-445c-be10-c84fa665caa4"}]: dispatch 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: audit 2026-03-07T06:55:05.870737+0000 mon.vm07 (mon.0) 365 : audit [INF] from='client.? 192.168.123.107:0/298761861' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "56cba12f-96c0-445c-be10-c84fa665caa4"}]: dispatch 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: audit 2026-03-07T06:55:05.873463+0000 mon.vm07 (mon.0) 366 : audit [INF] from='client.? 192.168.123.107:0/298761861' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "56cba12f-96c0-445c-be10-c84fa665caa4"}]': finished 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: audit 2026-03-07T06:55:05.873463+0000 mon.vm07 (mon.0) 366 : audit [INF] from='client.? 192.168.123.107:0/298761861' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "56cba12f-96c0-445c-be10-c84fa665caa4"}]': finished 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: cluster 2026-03-07T06:55:05.876088+0000 mon.vm07 (mon.0) 367 : cluster [DBG] osdmap e7: 2 total, 0 up, 2 in 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: cluster 2026-03-07T06:55:05.876088+0000 mon.vm07 (mon.0) 367 : cluster [DBG] osdmap e7: 2 total, 0 up, 2 in 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: audit 2026-03-07T06:55:05.876234+0000 mon.vm07 (mon.0) 368 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: audit 2026-03-07T06:55:05.876234+0000 mon.vm07 (mon.0) 368 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: audit 2026-03-07T06:55:05.876297+0000 mon.vm07 (mon.0) 369 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:06 vm09 bash[21220]: audit 2026-03-07T06:55:05.876297+0000 mon.vm07 (mon.0) 369 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:07.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:07 vm09 bash[21220]: audit 2026-03-07T06:55:06.436355+0000 mon.vm09 (mon.1) 3 : audit [DBG] from='client.? 192.168.123.109:0/1332580693' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:07.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:07 vm09 bash[21220]: audit 2026-03-07T06:55:06.436355+0000 mon.vm09 (mon.1) 3 : audit [DBG] from='client.? 192.168.123.109:0/1332580693' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:07.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:07 vm09 bash[21220]: audit 2026-03-07T06:55:06.492532+0000 mon.vm07 (mon.0) 370 : audit [DBG] from='client.? 192.168.123.107:0/685885353' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:07.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:07 vm09 bash[21220]: audit 2026-03-07T06:55:06.492532+0000 mon.vm07 (mon.0) 370 : audit [DBG] from='client.? 192.168.123.107:0/685885353' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:07.661 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:07 vm07 bash[17031]: audit 2026-03-07T06:55:06.436355+0000 mon.vm09 (mon.1) 3 : audit [DBG] from='client.? 192.168.123.109:0/1332580693' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:07.661 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:07 vm07 bash[17031]: audit 2026-03-07T06:55:06.436355+0000 mon.vm09 (mon.1) 3 : audit [DBG] from='client.? 192.168.123.109:0/1332580693' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:07.661 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:07 vm07 bash[17031]: audit 2026-03-07T06:55:06.492532+0000 mon.vm07 (mon.0) 370 : audit [DBG] from='client.? 192.168.123.107:0/685885353' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:07.661 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:07 vm07 bash[17031]: audit 2026-03-07T06:55:06.492532+0000 mon.vm07 (mon.0) 370 : audit [DBG] from='client.? 192.168.123.107:0/685885353' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:08.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:08 vm09 bash[21220]: cluster 2026-03-07T06:55:07.081778+0000 mgr.vm07.yrfcuj (mgr.14201) 66 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:08.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:08 vm09 bash[21220]: cluster 2026-03-07T06:55:07.081778+0000 mgr.vm07.yrfcuj (mgr.14201) 66 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:08.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:08 vm07 bash[17031]: cluster 2026-03-07T06:55:07.081778+0000 mgr.vm07.yrfcuj (mgr.14201) 66 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:08.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:08 vm07 bash[17031]: cluster 2026-03-07T06:55:07.081778+0000 mgr.vm07.yrfcuj (mgr.14201) 66 : cluster [DBG] pgmap v16: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:10.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:10 vm07 bash[17031]: cluster 2026-03-07T06:55:09.082026+0000 mgr.vm07.yrfcuj (mgr.14201) 67 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:10.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:10 vm07 bash[17031]: cluster 2026-03-07T06:55:09.082026+0000 mgr.vm07.yrfcuj (mgr.14201) 67 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:10.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:10 vm07 bash[17031]: audit 2026-03-07T06:55:10.095524+0000 mon.vm07 (mon.0) 371 : audit [INF] from='client.? 192.168.123.107:0/3373602686' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2a313541-d79f-4e9b-8bb5-038ec51951db"}]: dispatch 2026-03-07T07:55:10.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:10 vm07 bash[17031]: audit 2026-03-07T06:55:10.095524+0000 mon.vm07 (mon.0) 371 : audit [INF] from='client.? 192.168.123.107:0/3373602686' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2a313541-d79f-4e9b-8bb5-038ec51951db"}]: dispatch 2026-03-07T07:55:10.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:10 vm07 bash[17031]: audit 2026-03-07T06:55:10.107483+0000 mon.vm07 (mon.0) 372 : audit [INF] from='client.? 192.168.123.107:0/3373602686' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2a313541-d79f-4e9b-8bb5-038ec51951db"}]': finished 2026-03-07T07:55:10.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:10 vm07 bash[17031]: audit 2026-03-07T06:55:10.107483+0000 mon.vm07 (mon.0) 372 : audit [INF] from='client.? 192.168.123.107:0/3373602686' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2a313541-d79f-4e9b-8bb5-038ec51951db"}]': finished 2026-03-07T07:55:10.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:10 vm07 bash[17031]: cluster 2026-03-07T06:55:10.109633+0000 mon.vm07 (mon.0) 373 : cluster [DBG] osdmap e8: 3 total, 0 up, 3 in 2026-03-07T07:55:10.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:10 vm07 bash[17031]: cluster 2026-03-07T06:55:10.109633+0000 mon.vm07 (mon.0) 373 : cluster [DBG] osdmap e8: 3 total, 0 up, 3 in 2026-03-07T07:55:10.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:10 vm07 bash[17031]: audit 2026-03-07T06:55:10.109773+0000 mon.vm07 (mon.0) 374 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:10.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:10 vm07 bash[17031]: audit 2026-03-07T06:55:10.109773+0000 mon.vm07 (mon.0) 374 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:10.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:10 vm07 bash[17031]: audit 2026-03-07T06:55:10.109892+0000 mon.vm07 (mon.0) 375 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:10.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:10 vm07 bash[17031]: audit 2026-03-07T06:55:10.109892+0000 mon.vm07 (mon.0) 375 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:10.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:10 vm07 bash[17031]: audit 2026-03-07T06:55:10.109991+0000 mon.vm07 (mon.0) 376 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:10.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:10 vm07 bash[17031]: audit 2026-03-07T06:55:10.109991+0000 mon.vm07 (mon.0) 376 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:10.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:10 vm07 bash[17031]: audit 2026-03-07T06:55:10.124491+0000 mon.vm07 (mon.0) 377 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:55:10.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:10 vm07 bash[17031]: audit 2026-03-07T06:55:10.124491+0000 mon.vm07 (mon.0) 377 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:55:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:10 vm09 bash[21220]: cluster 2026-03-07T06:55:09.082026+0000 mgr.vm07.yrfcuj (mgr.14201) 67 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:10 vm09 bash[21220]: cluster 2026-03-07T06:55:09.082026+0000 mgr.vm07.yrfcuj (mgr.14201) 67 : cluster [DBG] pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:10 vm09 bash[21220]: audit 2026-03-07T06:55:10.095524+0000 mon.vm07 (mon.0) 371 : audit [INF] from='client.? 192.168.123.107:0/3373602686' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2a313541-d79f-4e9b-8bb5-038ec51951db"}]: dispatch 2026-03-07T07:55:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:10 vm09 bash[21220]: audit 2026-03-07T06:55:10.095524+0000 mon.vm07 (mon.0) 371 : audit [INF] from='client.? 192.168.123.107:0/3373602686' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2a313541-d79f-4e9b-8bb5-038ec51951db"}]: dispatch 2026-03-07T07:55:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:10 vm09 bash[21220]: audit 2026-03-07T06:55:10.107483+0000 mon.vm07 (mon.0) 372 : audit [INF] from='client.? 192.168.123.107:0/3373602686' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2a313541-d79f-4e9b-8bb5-038ec51951db"}]': finished 2026-03-07T07:55:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:10 vm09 bash[21220]: audit 2026-03-07T06:55:10.107483+0000 mon.vm07 (mon.0) 372 : audit [INF] from='client.? 192.168.123.107:0/3373602686' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2a313541-d79f-4e9b-8bb5-038ec51951db"}]': finished 2026-03-07T07:55:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:10 vm09 bash[21220]: cluster 2026-03-07T06:55:10.109633+0000 mon.vm07 (mon.0) 373 : cluster [DBG] osdmap e8: 3 total, 0 up, 3 in 2026-03-07T07:55:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:10 vm09 bash[21220]: cluster 2026-03-07T06:55:10.109633+0000 mon.vm07 (mon.0) 373 : cluster [DBG] osdmap e8: 3 total, 0 up, 3 in 2026-03-07T07:55:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:10 vm09 bash[21220]: audit 2026-03-07T06:55:10.109773+0000 mon.vm07 (mon.0) 374 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:10 vm09 bash[21220]: audit 2026-03-07T06:55:10.109773+0000 mon.vm07 (mon.0) 374 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:10 vm09 bash[21220]: audit 2026-03-07T06:55:10.109892+0000 mon.vm07 (mon.0) 375 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:10 vm09 bash[21220]: audit 2026-03-07T06:55:10.109892+0000 mon.vm07 (mon.0) 375 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:10 vm09 bash[21220]: audit 2026-03-07T06:55:10.109991+0000 mon.vm07 (mon.0) 376 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:10 vm09 bash[21220]: audit 2026-03-07T06:55:10.109991+0000 mon.vm07 (mon.0) 376 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:10 vm09 bash[21220]: audit 2026-03-07T06:55:10.124491+0000 mon.vm07 (mon.0) 377 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:55:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:10 vm09 bash[21220]: audit 2026-03-07T06:55:10.124491+0000 mon.vm07 (mon.0) 377 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:55:10.749 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:55:11.104 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:55:11.162 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":9,"num_osds":4,"num_up_osds":0,"osd_up_since":0,"num_in_osds":4,"osd_in_since":1772866510,"num_remapped_pgs":0} 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:10.283059+0000 mon.vm07 (mon.0) 378 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a583cddd-35d0-41d0-83c4-877b3c089e8b"}]: dispatch 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:10.283059+0000 mon.vm07 (mon.0) 378 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a583cddd-35d0-41d0-83c4-877b3c089e8b"}]: dispatch 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:10.285254+0000 mon.vm07 (mon.0) 379 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a583cddd-35d0-41d0-83c4-877b3c089e8b"}]': finished 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:10.285254+0000 mon.vm07 (mon.0) 379 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a583cddd-35d0-41d0-83c4-877b3c089e8b"}]': finished 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: cluster 2026-03-07T06:55:10.287333+0000 mon.vm07 (mon.0) 380 : cluster [DBG] osdmap e9: 4 total, 0 up, 4 in 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: cluster 2026-03-07T06:55:10.287333+0000 mon.vm07 (mon.0) 380 : cluster [DBG] osdmap e9: 4 total, 0 up, 4 in 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:10.287429+0000 mon.vm07 (mon.0) 381 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:10.287429+0000 mon.vm07 (mon.0) 381 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:10.287534+0000 mon.vm07 (mon.0) 382 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:10.287534+0000 mon.vm07 (mon.0) 382 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:10.287608+0000 mon.vm07 (mon.0) 383 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:10.287608+0000 mon.vm07 (mon.0) 383 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:10.287678+0000 mon.vm07 (mon.0) 384 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:10.287678+0000 mon.vm07 (mon.0) 384 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:10.287767+0000 mon.vm09 (mon.1) 4 : audit [INF] from='client.? 192.168.123.109:0/146919746' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a583cddd-35d0-41d0-83c4-877b3c089e8b"}]: dispatch 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:10.287767+0000 mon.vm09 (mon.1) 4 : audit [INF] from='client.? 192.168.123.109:0/146919746' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a583cddd-35d0-41d0-83c4-877b3c089e8b"}]: dispatch 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:10.707436+0000 mon.vm07 (mon.0) 385 : audit [DBG] from='client.? 192.168.123.107:0/1472318576' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:10.707436+0000 mon.vm07 (mon.0) 385 : audit [DBG] from='client.? 192.168.123.107:0/1472318576' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:10.889550+0000 mon.vm09 (mon.1) 5 : audit [DBG] from='client.? 192.168.123.109:0/3945368893' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:10.889550+0000 mon.vm09 (mon.1) 5 : audit [DBG] from='client.? 192.168.123.109:0/3945368893' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:11.098726+0000 mon.vm07 (mon.0) 386 : audit [DBG] from='client.? 192.168.123.107:0/1070274694' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:11.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:11 vm07 bash[17031]: audit 2026-03-07T06:55:11.098726+0000 mon.vm07 (mon.0) 386 : audit [DBG] from='client.? 192.168.123.107:0/1070274694' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:11.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:10.283059+0000 mon.vm07 (mon.0) 378 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a583cddd-35d0-41d0-83c4-877b3c089e8b"}]: dispatch 2026-03-07T07:55:11.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:10.283059+0000 mon.vm07 (mon.0) 378 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a583cddd-35d0-41d0-83c4-877b3c089e8b"}]: dispatch 2026-03-07T07:55:11.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:10.285254+0000 mon.vm07 (mon.0) 379 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a583cddd-35d0-41d0-83c4-877b3c089e8b"}]': finished 2026-03-07T07:55:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:10.285254+0000 mon.vm07 (mon.0) 379 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a583cddd-35d0-41d0-83c4-877b3c089e8b"}]': finished 2026-03-07T07:55:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: cluster 2026-03-07T06:55:10.287333+0000 mon.vm07 (mon.0) 380 : cluster [DBG] osdmap e9: 4 total, 0 up, 4 in 2026-03-07T07:55:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: cluster 2026-03-07T06:55:10.287333+0000 mon.vm07 (mon.0) 380 : cluster [DBG] osdmap e9: 4 total, 0 up, 4 in 2026-03-07T07:55:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:10.287429+0000 mon.vm07 (mon.0) 381 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:10.287429+0000 mon.vm07 (mon.0) 381 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:10.287534+0000 mon.vm07 (mon.0) 382 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:10.287534+0000 mon.vm07 (mon.0) 382 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:10.287608+0000 mon.vm07 (mon.0) 383 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:10.287608+0000 mon.vm07 (mon.0) 383 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:10.287678+0000 mon.vm07 (mon.0) 384 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:10.287678+0000 mon.vm07 (mon.0) 384 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:10.287767+0000 mon.vm09 (mon.1) 4 : audit [INF] from='client.? 192.168.123.109:0/146919746' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a583cddd-35d0-41d0-83c4-877b3c089e8b"}]: dispatch 2026-03-07T07:55:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:10.287767+0000 mon.vm09 (mon.1) 4 : audit [INF] from='client.? 192.168.123.109:0/146919746' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a583cddd-35d0-41d0-83c4-877b3c089e8b"}]: dispatch 2026-03-07T07:55:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:10.707436+0000 mon.vm07 (mon.0) 385 : audit [DBG] from='client.? 192.168.123.107:0/1472318576' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:10.707436+0000 mon.vm07 (mon.0) 385 : audit [DBG] from='client.? 192.168.123.107:0/1472318576' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:10.889550+0000 mon.vm09 (mon.1) 5 : audit [DBG] from='client.? 192.168.123.109:0/3945368893' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:10.889550+0000 mon.vm09 (mon.1) 5 : audit [DBG] from='client.? 192.168.123.109:0/3945368893' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:11.098726+0000 mon.vm07 (mon.0) 386 : audit [DBG] from='client.? 192.168.123.107:0/1070274694' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:11 vm09 bash[21220]: audit 2026-03-07T06:55:11.098726+0000 mon.vm07 (mon.0) 386 : audit [DBG] from='client.? 192.168.123.107:0/1070274694' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:12.162 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd stat -f json 2026-03-07T07:55:12.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:12 vm09 bash[21220]: cluster 2026-03-07T06:55:11.082214+0000 mgr.vm07.yrfcuj (mgr.14201) 68 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:12.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:12 vm09 bash[21220]: cluster 2026-03-07T06:55:11.082214+0000 mgr.vm07.yrfcuj (mgr.14201) 68 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:12.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:12 vm07 bash[17031]: cluster 2026-03-07T06:55:11.082214+0000 mgr.vm07.yrfcuj (mgr.14201) 68 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:12.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:12 vm07 bash[17031]: cluster 2026-03-07T06:55:11.082214+0000 mgr.vm07.yrfcuj (mgr.14201) 68 : cluster [DBG] pgmap v20: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:14.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: cluster 2026-03-07T06:55:13.082451+0000 mgr.vm07.yrfcuj (mgr.14201) 69 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:14.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: cluster 2026-03-07T06:55:13.082451+0000 mgr.vm07.yrfcuj (mgr.14201) 69 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:14.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: audit 2026-03-07T06:55:14.206094+0000 mon.vm07 (mon.0) 387 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2f651ee5-5a7b-4d0c-82f4-907f0a8beddc"}]: dispatch 2026-03-07T07:55:14.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: audit 2026-03-07T06:55:14.206094+0000 mon.vm07 (mon.0) 387 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2f651ee5-5a7b-4d0c-82f4-907f0a8beddc"}]: dispatch 2026-03-07T07:55:14.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: audit 2026-03-07T06:55:14.208749+0000 mon.vm07 (mon.0) 388 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2f651ee5-5a7b-4d0c-82f4-907f0a8beddc"}]': finished 2026-03-07T07:55:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: audit 2026-03-07T06:55:14.208749+0000 mon.vm07 (mon.0) 388 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2f651ee5-5a7b-4d0c-82f4-907f0a8beddc"}]': finished 2026-03-07T07:55:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: audit 2026-03-07T06:55:14.210746+0000 mon.vm09 (mon.1) 6 : audit [INF] from='client.? 192.168.123.109:0/3480598887' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2f651ee5-5a7b-4d0c-82f4-907f0a8beddc"}]: dispatch 2026-03-07T07:55:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: audit 2026-03-07T06:55:14.210746+0000 mon.vm09 (mon.1) 6 : audit [INF] from='client.? 192.168.123.109:0/3480598887' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2f651ee5-5a7b-4d0c-82f4-907f0a8beddc"}]: dispatch 2026-03-07T07:55:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: cluster 2026-03-07T06:55:14.211646+0000 mon.vm07 (mon.0) 389 : cluster [DBG] osdmap e10: 5 total, 0 up, 5 in 2026-03-07T07:55:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: cluster 2026-03-07T06:55:14.211646+0000 mon.vm07 (mon.0) 389 : cluster [DBG] osdmap e10: 5 total, 0 up, 5 in 2026-03-07T07:55:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: audit 2026-03-07T06:55:14.211847+0000 mon.vm07 (mon.0) 390 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: audit 2026-03-07T06:55:14.211847+0000 mon.vm07 (mon.0) 390 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: audit 2026-03-07T06:55:14.211980+0000 mon.vm07 (mon.0) 391 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: audit 2026-03-07T06:55:14.211980+0000 mon.vm07 (mon.0) 391 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: audit 2026-03-07T06:55:14.212092+0000 mon.vm07 (mon.0) 392 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: audit 2026-03-07T06:55:14.212092+0000 mon.vm07 (mon.0) 392 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: audit 2026-03-07T06:55:14.212205+0000 mon.vm07 (mon.0) 393 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: audit 2026-03-07T06:55:14.212205+0000 mon.vm07 (mon.0) 393 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: audit 2026-03-07T06:55:14.212313+0000 mon.vm07 (mon.0) 394 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:14 vm09 bash[21220]: audit 2026-03-07T06:55:14.212313+0000 mon.vm07 (mon.0) 394 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: cluster 2026-03-07T06:55:13.082451+0000 mgr.vm07.yrfcuj (mgr.14201) 69 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: cluster 2026-03-07T06:55:13.082451+0000 mgr.vm07.yrfcuj (mgr.14201) 69 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: audit 2026-03-07T06:55:14.206094+0000 mon.vm07 (mon.0) 387 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2f651ee5-5a7b-4d0c-82f4-907f0a8beddc"}]: dispatch 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: audit 2026-03-07T06:55:14.206094+0000 mon.vm07 (mon.0) 387 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2f651ee5-5a7b-4d0c-82f4-907f0a8beddc"}]: dispatch 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: audit 2026-03-07T06:55:14.208749+0000 mon.vm07 (mon.0) 388 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2f651ee5-5a7b-4d0c-82f4-907f0a8beddc"}]': finished 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: audit 2026-03-07T06:55:14.208749+0000 mon.vm07 (mon.0) 388 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2f651ee5-5a7b-4d0c-82f4-907f0a8beddc"}]': finished 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: audit 2026-03-07T06:55:14.210746+0000 mon.vm09 (mon.1) 6 : audit [INF] from='client.? 192.168.123.109:0/3480598887' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2f651ee5-5a7b-4d0c-82f4-907f0a8beddc"}]: dispatch 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: audit 2026-03-07T06:55:14.210746+0000 mon.vm09 (mon.1) 6 : audit [INF] from='client.? 192.168.123.109:0/3480598887' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "2f651ee5-5a7b-4d0c-82f4-907f0a8beddc"}]: dispatch 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: cluster 2026-03-07T06:55:14.211646+0000 mon.vm07 (mon.0) 389 : cluster [DBG] osdmap e10: 5 total, 0 up, 5 in 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: cluster 2026-03-07T06:55:14.211646+0000 mon.vm07 (mon.0) 389 : cluster [DBG] osdmap e10: 5 total, 0 up, 5 in 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: audit 2026-03-07T06:55:14.211847+0000 mon.vm07 (mon.0) 390 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: audit 2026-03-07T06:55:14.211847+0000 mon.vm07 (mon.0) 390 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: audit 2026-03-07T06:55:14.211980+0000 mon.vm07 (mon.0) 391 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: audit 2026-03-07T06:55:14.211980+0000 mon.vm07 (mon.0) 391 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: audit 2026-03-07T06:55:14.212092+0000 mon.vm07 (mon.0) 392 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: audit 2026-03-07T06:55:14.212092+0000 mon.vm07 (mon.0) 392 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: audit 2026-03-07T06:55:14.212205+0000 mon.vm07 (mon.0) 393 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: audit 2026-03-07T06:55:14.212205+0000 mon.vm07 (mon.0) 393 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: audit 2026-03-07T06:55:14.212313+0000 mon.vm07 (mon.0) 394 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:14.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:14 vm07 bash[17031]: audit 2026-03-07T06:55:14.212313+0000 mon.vm07 (mon.0) 394 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:15.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:14.419868+0000 mon.vm07 (mon.0) 395 : audit [INF] from='client.? 192.168.123.107:0/3508554448' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5874e093-78e8-4710-8cf9-d27b73d59b9b"}]: dispatch 2026-03-07T07:55:15.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:14.419868+0000 mon.vm07 (mon.0) 395 : audit [INF] from='client.? 192.168.123.107:0/3508554448' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5874e093-78e8-4710-8cf9-d27b73d59b9b"}]: dispatch 2026-03-07T07:55:15.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:14.422923+0000 mon.vm07 (mon.0) 396 : audit [INF] from='client.? 192.168.123.107:0/3508554448' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5874e093-78e8-4710-8cf9-d27b73d59b9b"}]': finished 2026-03-07T07:55:15.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:14.422923+0000 mon.vm07 (mon.0) 396 : audit [INF] from='client.? 192.168.123.107:0/3508554448' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5874e093-78e8-4710-8cf9-d27b73d59b9b"}]': finished 2026-03-07T07:55:15.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: cluster 2026-03-07T06:55:14.425331+0000 mon.vm07 (mon.0) 397 : cluster [DBG] osdmap e11: 6 total, 0 up, 6 in 2026-03-07T07:55:15.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: cluster 2026-03-07T06:55:14.425331+0000 mon.vm07 (mon.0) 397 : cluster [DBG] osdmap e11: 6 total, 0 up, 6 in 2026-03-07T07:55:15.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:14.425553+0000 mon.vm07 (mon.0) 398 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:15.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:14.425553+0000 mon.vm07 (mon.0) 398 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:15.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:14.425683+0000 mon.vm07 (mon.0) 399 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:15.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:14.425683+0000 mon.vm07 (mon.0) 399 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:15.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:14.425768+0000 mon.vm07 (mon.0) 400 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:15.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:14.425768+0000 mon.vm07 (mon.0) 400 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:15.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:14.425862+0000 mon.vm07 (mon.0) 401 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:15.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:14.425862+0000 mon.vm07 (mon.0) 401 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:15.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:14.425944+0000 mon.vm07 (mon.0) 402 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:15.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:14.425944+0000 mon.vm07 (mon.0) 402 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:15.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:14.426030+0000 mon.vm07 (mon.0) 403 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:15.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:14.426030+0000 mon.vm07 (mon.0) 403 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:15.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:14.771884+0000 mon.vm09 (mon.1) 7 : audit [DBG] from='client.? 192.168.123.109:0/653768529' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:15.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:14.771884+0000 mon.vm09 (mon.1) 7 : audit [DBG] from='client.? 192.168.123.109:0/653768529' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:15.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:15.006014+0000 mon.vm07 (mon.0) 404 : audit [DBG] from='client.? 192.168.123.107:0/1888682018' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:15.409 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:15 vm07 bash[17031]: audit 2026-03-07T06:55:15.006014+0000 mon.vm07 (mon.0) 404 : audit [DBG] from='client.? 192.168.123.107:0/1888682018' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:15.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:14.419868+0000 mon.vm07 (mon.0) 395 : audit [INF] from='client.? 192.168.123.107:0/3508554448' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5874e093-78e8-4710-8cf9-d27b73d59b9b"}]: dispatch 2026-03-07T07:55:15.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:14.419868+0000 mon.vm07 (mon.0) 395 : audit [INF] from='client.? 192.168.123.107:0/3508554448' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "5874e093-78e8-4710-8cf9-d27b73d59b9b"}]: dispatch 2026-03-07T07:55:15.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:14.422923+0000 mon.vm07 (mon.0) 396 : audit [INF] from='client.? 192.168.123.107:0/3508554448' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5874e093-78e8-4710-8cf9-d27b73d59b9b"}]': finished 2026-03-07T07:55:15.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:14.422923+0000 mon.vm07 (mon.0) 396 : audit [INF] from='client.? 192.168.123.107:0/3508554448' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5874e093-78e8-4710-8cf9-d27b73d59b9b"}]': finished 2026-03-07T07:55:15.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: cluster 2026-03-07T06:55:14.425331+0000 mon.vm07 (mon.0) 397 : cluster [DBG] osdmap e11: 6 total, 0 up, 6 in 2026-03-07T07:55:15.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: cluster 2026-03-07T06:55:14.425331+0000 mon.vm07 (mon.0) 397 : cluster [DBG] osdmap e11: 6 total, 0 up, 6 in 2026-03-07T07:55:15.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:14.425553+0000 mon.vm07 (mon.0) 398 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:14.425553+0000 mon.vm07 (mon.0) 398 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:14.425683+0000 mon.vm07 (mon.0) 399 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:14.425683+0000 mon.vm07 (mon.0) 399 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:14.425768+0000 mon.vm07 (mon.0) 400 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:14.425768+0000 mon.vm07 (mon.0) 400 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:14.425862+0000 mon.vm07 (mon.0) 401 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:14.425862+0000 mon.vm07 (mon.0) 401 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:14.425944+0000 mon.vm07 (mon.0) 402 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:14.425944+0000 mon.vm07 (mon.0) 402 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:14.426030+0000 mon.vm07 (mon.0) 403 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:14.426030+0000 mon.vm07 (mon.0) 403 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:14.771884+0000 mon.vm09 (mon.1) 7 : audit [DBG] from='client.? 192.168.123.109:0/653768529' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:14.771884+0000 mon.vm09 (mon.1) 7 : audit [DBG] from='client.? 192.168.123.109:0/653768529' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:15.006014+0000 mon.vm07 (mon.0) 404 : audit [DBG] from='client.? 192.168.123.107:0/1888682018' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:15 vm09 bash[21220]: audit 2026-03-07T06:55:15.006014+0000 mon.vm07 (mon.0) 404 : audit [DBG] from='client.? 192.168.123.107:0/1888682018' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:16.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:16 vm09 bash[21220]: cluster 2026-03-07T06:55:15.082637+0000 mgr.vm07.yrfcuj (mgr.14201) 70 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:16.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:16 vm09 bash[21220]: cluster 2026-03-07T06:55:15.082637+0000 mgr.vm07.yrfcuj (mgr.14201) 70 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:16.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:16 vm07 bash[17031]: cluster 2026-03-07T06:55:15.082637+0000 mgr.vm07.yrfcuj (mgr.14201) 70 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:16.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:16 vm07 bash[17031]: cluster 2026-03-07T06:55:15.082637+0000 mgr.vm07.yrfcuj (mgr.14201) 70 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:16.930 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:55:17.334 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:55:17.406 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":11,"num_osds":6,"num_up_osds":0,"osd_up_since":0,"num_in_osds":6,"osd_in_since":1772866514,"num_remapped_pgs":0} 2026-03-07T07:55:17.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:17 vm07 bash[17031]: audit 2026-03-07T06:55:17.328739+0000 mon.vm07 (mon.0) 405 : audit [DBG] from='client.? 192.168.123.107:0/2292850326' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:17.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:17 vm07 bash[17031]: audit 2026-03-07T06:55:17.328739+0000 mon.vm07 (mon.0) 405 : audit [DBG] from='client.? 192.168.123.107:0/2292850326' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:17.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:17 vm09 bash[21220]: audit 2026-03-07T06:55:17.328739+0000 mon.vm07 (mon.0) 405 : audit [DBG] from='client.? 192.168.123.107:0/2292850326' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:17.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:17 vm09 bash[21220]: audit 2026-03-07T06:55:17.328739+0000 mon.vm07 (mon.0) 405 : audit [DBG] from='client.? 192.168.123.107:0/2292850326' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:18.407 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd stat -f json 2026-03-07T07:55:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: cluster 2026-03-07T06:55:17.082824+0000 mgr.vm07.yrfcuj (mgr.14201) 71 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: cluster 2026-03-07T06:55:17.082824+0000 mgr.vm07.yrfcuj (mgr.14201) 71 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.133246+0000 mon.vm07 (mon.0) 406 : audit [INF] from='client.? 192.168.123.107:0/1125204967' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "85f57ef2-7375-4150-b632-6d08ce1f021e"}]: dispatch 2026-03-07T07:55:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.133246+0000 mon.vm07 (mon.0) 406 : audit [INF] from='client.? 192.168.123.107:0/1125204967' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "85f57ef2-7375-4150-b632-6d08ce1f021e"}]: dispatch 2026-03-07T07:55:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.136591+0000 mon.vm07 (mon.0) 407 : audit [INF] from='client.? 192.168.123.107:0/1125204967' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "85f57ef2-7375-4150-b632-6d08ce1f021e"}]': finished 2026-03-07T07:55:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.136591+0000 mon.vm07 (mon.0) 407 : audit [INF] from='client.? 192.168.123.107:0/1125204967' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "85f57ef2-7375-4150-b632-6d08ce1f021e"}]': finished 2026-03-07T07:55:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: cluster 2026-03-07T06:55:18.138486+0000 mon.vm07 (mon.0) 408 : cluster [DBG] osdmap e12: 7 total, 0 up, 7 in 2026-03-07T07:55:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: cluster 2026-03-07T06:55:18.138486+0000 mon.vm07 (mon.0) 408 : cluster [DBG] osdmap e12: 7 total, 0 up, 7 in 2026-03-07T07:55:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.139212+0000 mon.vm07 (mon.0) 409 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.139212+0000 mon.vm07 (mon.0) 409 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.139574+0000 mon.vm07 (mon.0) 410 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.139574+0000 mon.vm07 (mon.0) 410 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.139921+0000 mon.vm07 (mon.0) 411 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.139921+0000 mon.vm07 (mon.0) 411 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.140266+0000 mon.vm07 (mon.0) 412 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.140266+0000 mon.vm07 (mon.0) 412 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.140605+0000 mon.vm07 (mon.0) 413 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.140605+0000 mon.vm07 (mon.0) 413 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.140949+0000 mon.vm07 (mon.0) 414 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.140949+0000 mon.vm07 (mon.0) 414 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.141364+0000 mon.vm07 (mon.0) 415 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.141364+0000 mon.vm07 (mon.0) 415 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.144949+0000 mon.vm07 (mon.0) 416 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6ea2bf16-f584-49a8-ba45-a4b9ba5043ce"}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.144949+0000 mon.vm07 (mon.0) 416 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6ea2bf16-f584-49a8-ba45-a4b9ba5043ce"}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.147883+0000 mon.vm07 (mon.0) 417 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "6ea2bf16-f584-49a8-ba45-a4b9ba5043ce"}]': finished 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.147883+0000 mon.vm07 (mon.0) 417 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "6ea2bf16-f584-49a8-ba45-a4b9ba5043ce"}]': finished 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.149135+0000 mon.vm09 (mon.1) 8 : audit [INF] from='client.? 192.168.123.109:0/3908708691' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6ea2bf16-f584-49a8-ba45-a4b9ba5043ce"}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.149135+0000 mon.vm09 (mon.1) 8 : audit [INF] from='client.? 192.168.123.109:0/3908708691' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6ea2bf16-f584-49a8-ba45-a4b9ba5043ce"}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: cluster 2026-03-07T06:55:18.150024+0000 mon.vm07 (mon.0) 418 : cluster [DBG] osdmap e13: 8 total, 0 up, 8 in 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: cluster 2026-03-07T06:55:18.150024+0000 mon.vm07 (mon.0) 418 : cluster [DBG] osdmap e13: 8 total, 0 up, 8 in 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.150679+0000 mon.vm07 (mon.0) 419 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.150679+0000 mon.vm07 (mon.0) 419 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.151074+0000 mon.vm07 (mon.0) 420 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.151074+0000 mon.vm07 (mon.0) 420 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.151437+0000 mon.vm07 (mon.0) 421 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.151437+0000 mon.vm07 (mon.0) 421 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.151776+0000 mon.vm07 (mon.0) 422 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.151776+0000 mon.vm07 (mon.0) 422 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.152075+0000 mon.vm07 (mon.0) 423 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.152075+0000 mon.vm07 (mon.0) 423 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.152544+0000 mon.vm07 (mon.0) 424 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.152544+0000 mon.vm07 (mon.0) 424 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.152912+0000 mon.vm07 (mon.0) 425 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.152912+0000 mon.vm07 (mon.0) 425 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.154958+0000 mon.vm07 (mon.0) 426 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:18.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:18 vm07 bash[17031]: audit 2026-03-07T06:55:18.154958+0000 mon.vm07 (mon.0) 426 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:18.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: cluster 2026-03-07T06:55:17.082824+0000 mgr.vm07.yrfcuj (mgr.14201) 71 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: cluster 2026-03-07T06:55:17.082824+0000 mgr.vm07.yrfcuj (mgr.14201) 71 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.133246+0000 mon.vm07 (mon.0) 406 : audit [INF] from='client.? 192.168.123.107:0/1125204967' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "85f57ef2-7375-4150-b632-6d08ce1f021e"}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.133246+0000 mon.vm07 (mon.0) 406 : audit [INF] from='client.? 192.168.123.107:0/1125204967' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "85f57ef2-7375-4150-b632-6d08ce1f021e"}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.136591+0000 mon.vm07 (mon.0) 407 : audit [INF] from='client.? 192.168.123.107:0/1125204967' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "85f57ef2-7375-4150-b632-6d08ce1f021e"}]': finished 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.136591+0000 mon.vm07 (mon.0) 407 : audit [INF] from='client.? 192.168.123.107:0/1125204967' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "85f57ef2-7375-4150-b632-6d08ce1f021e"}]': finished 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: cluster 2026-03-07T06:55:18.138486+0000 mon.vm07 (mon.0) 408 : cluster [DBG] osdmap e12: 7 total, 0 up, 7 in 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: cluster 2026-03-07T06:55:18.138486+0000 mon.vm07 (mon.0) 408 : cluster [DBG] osdmap e12: 7 total, 0 up, 7 in 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.139212+0000 mon.vm07 (mon.0) 409 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.139212+0000 mon.vm07 (mon.0) 409 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.139574+0000 mon.vm07 (mon.0) 410 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.139574+0000 mon.vm07 (mon.0) 410 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.139921+0000 mon.vm07 (mon.0) 411 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.139921+0000 mon.vm07 (mon.0) 411 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.140266+0000 mon.vm07 (mon.0) 412 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.140266+0000 mon.vm07 (mon.0) 412 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.140605+0000 mon.vm07 (mon.0) 413 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.140605+0000 mon.vm07 (mon.0) 413 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.140949+0000 mon.vm07 (mon.0) 414 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.140949+0000 mon.vm07 (mon.0) 414 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.141364+0000 mon.vm07 (mon.0) 415 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.141364+0000 mon.vm07 (mon.0) 415 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.144949+0000 mon.vm07 (mon.0) 416 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6ea2bf16-f584-49a8-ba45-a4b9ba5043ce"}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.144949+0000 mon.vm07 (mon.0) 416 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6ea2bf16-f584-49a8-ba45-a4b9ba5043ce"}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.147883+0000 mon.vm07 (mon.0) 417 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "6ea2bf16-f584-49a8-ba45-a4b9ba5043ce"}]': finished 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.147883+0000 mon.vm07 (mon.0) 417 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "6ea2bf16-f584-49a8-ba45-a4b9ba5043ce"}]': finished 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.149135+0000 mon.vm09 (mon.1) 8 : audit [INF] from='client.? 192.168.123.109:0/3908708691' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6ea2bf16-f584-49a8-ba45-a4b9ba5043ce"}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.149135+0000 mon.vm09 (mon.1) 8 : audit [INF] from='client.? 192.168.123.109:0/3908708691' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "6ea2bf16-f584-49a8-ba45-a4b9ba5043ce"}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: cluster 2026-03-07T06:55:18.150024+0000 mon.vm07 (mon.0) 418 : cluster [DBG] osdmap e13: 8 total, 0 up, 8 in 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: cluster 2026-03-07T06:55:18.150024+0000 mon.vm07 (mon.0) 418 : cluster [DBG] osdmap e13: 8 total, 0 up, 8 in 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.150679+0000 mon.vm07 (mon.0) 419 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.150679+0000 mon.vm07 (mon.0) 419 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.151074+0000 mon.vm07 (mon.0) 420 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.151074+0000 mon.vm07 (mon.0) 420 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.151437+0000 mon.vm07 (mon.0) 421 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.151437+0000 mon.vm07 (mon.0) 421 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.151776+0000 mon.vm07 (mon.0) 422 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.151776+0000 mon.vm07 (mon.0) 422 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.152075+0000 mon.vm07 (mon.0) 423 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.152075+0000 mon.vm07 (mon.0) 423 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.152544+0000 mon.vm07 (mon.0) 424 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.152544+0000 mon.vm07 (mon.0) 424 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.152912+0000 mon.vm07 (mon.0) 425 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.152912+0000 mon.vm07 (mon.0) 425 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.154958+0000 mon.vm07 (mon.0) 426 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:18 vm09 bash[21220]: audit 2026-03-07T06:55:18.154958+0000 mon.vm07 (mon.0) 426 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:19.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:19 vm09 bash[21220]: audit 2026-03-07T06:55:18.766795+0000 mon.vm09 (mon.1) 9 : audit [DBG] from='client.? 192.168.123.109:0/1815062261' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:19.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:19 vm09 bash[21220]: audit 2026-03-07T06:55:18.766795+0000 mon.vm09 (mon.1) 9 : audit [DBG] from='client.? 192.168.123.109:0/1815062261' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:19.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:19 vm09 bash[21220]: audit 2026-03-07T06:55:18.802719+0000 mon.vm07 (mon.0) 427 : audit [DBG] from='client.? 192.168.123.107:0/2443536930' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:19.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:19 vm09 bash[21220]: audit 2026-03-07T06:55:18.802719+0000 mon.vm07 (mon.0) 427 : audit [DBG] from='client.? 192.168.123.107:0/2443536930' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:19.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:19 vm07 bash[17031]: audit 2026-03-07T06:55:18.766795+0000 mon.vm09 (mon.1) 9 : audit [DBG] from='client.? 192.168.123.109:0/1815062261' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:19.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:19 vm07 bash[17031]: audit 2026-03-07T06:55:18.766795+0000 mon.vm09 (mon.1) 9 : audit [DBG] from='client.? 192.168.123.109:0/1815062261' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:19.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:19 vm07 bash[17031]: audit 2026-03-07T06:55:18.802719+0000 mon.vm07 (mon.0) 427 : audit [DBG] from='client.? 192.168.123.107:0/2443536930' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:19.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:19 vm07 bash[17031]: audit 2026-03-07T06:55:18.802719+0000 mon.vm07 (mon.0) 427 : audit [DBG] from='client.? 192.168.123.107:0/2443536930' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-07T07:55:20.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:20 vm09 bash[21220]: cluster 2026-03-07T06:55:19.083060+0000 mgr.vm07.yrfcuj (mgr.14201) 72 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:20.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:20 vm09 bash[21220]: cluster 2026-03-07T06:55:19.083060+0000 mgr.vm07.yrfcuj (mgr.14201) 72 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:20.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:20 vm07 bash[17031]: cluster 2026-03-07T06:55:19.083060+0000 mgr.vm07.yrfcuj (mgr.14201) 72 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:20.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:20 vm07 bash[17031]: cluster 2026-03-07T06:55:19.083060+0000 mgr.vm07.yrfcuj (mgr.14201) 72 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:22.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:22 vm09 bash[21220]: cluster 2026-03-07T06:55:21.083249+0000 mgr.vm07.yrfcuj (mgr.14201) 73 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:22.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:22 vm09 bash[21220]: cluster 2026-03-07T06:55:21.083249+0000 mgr.vm07.yrfcuj (mgr.14201) 73 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:22.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:22 vm07 bash[17031]: cluster 2026-03-07T06:55:21.083249+0000 mgr.vm07.yrfcuj (mgr.14201) 73 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:22.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:22 vm07 bash[17031]: cluster 2026-03-07T06:55:21.083249+0000 mgr.vm07.yrfcuj (mgr.14201) 73 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:23.199 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:55:23.531 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:55:23.603 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":13,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1772866518,"num_remapped_pgs":0} 2026-03-07T07:55:24.604 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd stat -f json 2026-03-07T07:55:24.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:24 vm09 bash[21220]: cluster 2026-03-07T06:55:23.083441+0000 mgr.vm07.yrfcuj (mgr.14201) 74 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:24.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:24 vm09 bash[21220]: cluster 2026-03-07T06:55:23.083441+0000 mgr.vm07.yrfcuj (mgr.14201) 74 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:24.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:24 vm09 bash[21220]: audit 2026-03-07T06:55:23.526343+0000 mon.vm07 (mon.0) 428 : audit [DBG] from='client.? 192.168.123.107:0/3730897975' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:24.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:24 vm09 bash[21220]: audit 2026-03-07T06:55:23.526343+0000 mon.vm07 (mon.0) 428 : audit [DBG] from='client.? 192.168.123.107:0/3730897975' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:24 vm07 bash[17031]: cluster 2026-03-07T06:55:23.083441+0000 mgr.vm07.yrfcuj (mgr.14201) 74 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:24 vm07 bash[17031]: cluster 2026-03-07T06:55:23.083441+0000 mgr.vm07.yrfcuj (mgr.14201) 74 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:24 vm07 bash[17031]: audit 2026-03-07T06:55:23.526343+0000 mon.vm07 (mon.0) 428 : audit [DBG] from='client.? 192.168.123.107:0/3730897975' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:24 vm07 bash[17031]: audit 2026-03-07T06:55:23.526343+0000 mon.vm07 (mon.0) 428 : audit [DBG] from='client.? 192.168.123.107:0/3730897975' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:25.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:25 vm09 bash[21220]: audit 2026-03-07T06:55:25.124627+0000 mon.vm07 (mon.0) 429 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:55:25.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:25 vm09 bash[21220]: audit 2026-03-07T06:55:25.124627+0000 mon.vm07 (mon.0) 429 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:55:25.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:25 vm07 bash[17031]: audit 2026-03-07T06:55:25.124627+0000 mon.vm07 (mon.0) 429 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:55:25.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:25 vm07 bash[17031]: audit 2026-03-07T06:55:25.124627+0000 mon.vm07 (mon.0) 429 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:55:26.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:26 vm07 bash[17031]: cluster 2026-03-07T06:55:25.083622+0000 mgr.vm07.yrfcuj (mgr.14201) 75 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:26.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:26 vm07 bash[17031]: cluster 2026-03-07T06:55:25.083622+0000 mgr.vm07.yrfcuj (mgr.14201) 75 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:26.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:26 vm09 bash[21220]: cluster 2026-03-07T06:55:25.083622+0000 mgr.vm07.yrfcuj (mgr.14201) 75 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:26.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:26 vm09 bash[21220]: cluster 2026-03-07T06:55:25.083622+0000 mgr.vm07.yrfcuj (mgr.14201) 75 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:27.760 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:27 vm07 bash[17031]: audit 2026-03-07T06:55:26.927160+0000 mon.vm07 (mon.0) 430 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-07T07:55:27.761 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:27 vm07 bash[17031]: audit 2026-03-07T06:55:26.927160+0000 mon.vm07 (mon.0) 430 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-07T07:55:27.761 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:27 vm07 bash[17031]: audit 2026-03-07T06:55:26.927620+0000 mon.vm07 (mon.0) 431 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:27.761 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:27 vm07 bash[17031]: audit 2026-03-07T06:55:26.927620+0000 mon.vm07 (mon.0) 431 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:27.761 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:27 vm07 bash[17031]: audit 2026-03-07T06:55:27.186352+0000 mon.vm07 (mon.0) 432 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-07T07:55:27.761 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:27 vm07 bash[17031]: audit 2026-03-07T06:55:27.186352+0000 mon.vm07 (mon.0) 432 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-07T07:55:27.761 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:27 vm07 bash[17031]: audit 2026-03-07T06:55:27.187079+0000 mon.vm07 (mon.0) 433 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:27.761 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:27 vm07 bash[17031]: audit 2026-03-07T06:55:27.187079+0000 mon.vm07 (mon.0) 433 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:27.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:27 vm09 bash[21220]: audit 2026-03-07T06:55:26.927160+0000 mon.vm07 (mon.0) 430 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-07T07:55:27.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:27 vm09 bash[21220]: audit 2026-03-07T06:55:26.927160+0000 mon.vm07 (mon.0) 430 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-07T07:55:27.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:27 vm09 bash[21220]: audit 2026-03-07T06:55:26.927620+0000 mon.vm07 (mon.0) 431 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:27.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:27 vm09 bash[21220]: audit 2026-03-07T06:55:26.927620+0000 mon.vm07 (mon.0) 431 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:27.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:27 vm09 bash[21220]: audit 2026-03-07T06:55:27.186352+0000 mon.vm07 (mon.0) 432 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-07T07:55:27.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:27 vm09 bash[21220]: audit 2026-03-07T06:55:27.186352+0000 mon.vm07 (mon.0) 432 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-07T07:55:27.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:27 vm09 bash[21220]: audit 2026-03-07T06:55:27.187079+0000 mon.vm07 (mon.0) 433 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:27.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:27 vm09 bash[21220]: audit 2026-03-07T06:55:27.187079+0000 mon.vm07 (mon.0) 433 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:28.036 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:27 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:55:28.261 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:28 vm09 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:55:28.313 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:28 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:55:28.547 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:28 vm09 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:55:28.547 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:28 vm09 bash[21220]: cephadm 2026-03-07T06:55:26.927987+0000 mgr.vm07.yrfcuj (mgr.14201) 76 : cephadm [INF] Deploying daemon osd.1 on vm07 2026-03-07T07:55:28.547 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:28 vm09 bash[21220]: cephadm 2026-03-07T06:55:26.927987+0000 mgr.vm07.yrfcuj (mgr.14201) 76 : cephadm [INF] Deploying daemon osd.1 on vm07 2026-03-07T07:55:28.547 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:28 vm09 bash[21220]: cluster 2026-03-07T06:55:27.083846+0000 mgr.vm07.yrfcuj (mgr.14201) 77 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:28.547 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:28 vm09 bash[21220]: cluster 2026-03-07T06:55:27.083846+0000 mgr.vm07.yrfcuj (mgr.14201) 77 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:28.547 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:28 vm09 bash[21220]: cephadm 2026-03-07T06:55:27.187478+0000 mgr.vm07.yrfcuj (mgr.14201) 78 : cephadm [INF] Deploying daemon osd.0 on vm09 2026-03-07T07:55:28.547 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:28 vm09 bash[21220]: cephadm 2026-03-07T06:55:27.187478+0000 mgr.vm07.yrfcuj (mgr.14201) 78 : cephadm [INF] Deploying daemon osd.0 on vm09 2026-03-07T07:55:28.547 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:28 vm09 bash[21220]: audit 2026-03-07T06:55:28.249707+0000 mon.vm07 (mon.0) 434 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:28.547 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:28 vm09 bash[21220]: audit 2026-03-07T06:55:28.249707+0000 mon.vm07 (mon.0) 434 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:28.547 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:28 vm09 bash[21220]: audit 2026-03-07T06:55:28.268274+0000 mon.vm07 (mon.0) 435 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:28.547 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:28 vm09 bash[21220]: audit 2026-03-07T06:55:28.268274+0000 mon.vm07 (mon.0) 435 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:28.547 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:28 vm09 bash[21220]: audit 2026-03-07T06:55:28.269225+0000 mon.vm07 (mon.0) 436 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-07T07:55:28.547 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:28 vm09 bash[21220]: audit 2026-03-07T06:55:28.269225+0000 mon.vm07 (mon.0) 436 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-07T07:55:28.547 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:28 vm09 bash[21220]: audit 2026-03-07T06:55:28.269773+0000 mon.vm07 (mon.0) 437 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:28.547 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:28 vm09 bash[21220]: audit 2026-03-07T06:55:28.269773+0000 mon.vm07 (mon.0) 437 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:28.636 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:28 vm07 bash[17031]: cephadm 2026-03-07T06:55:26.927987+0000 mgr.vm07.yrfcuj (mgr.14201) 76 : cephadm [INF] Deploying daemon osd.1 on vm07 2026-03-07T07:55:28.636 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:28 vm07 bash[17031]: cephadm 2026-03-07T06:55:26.927987+0000 mgr.vm07.yrfcuj (mgr.14201) 76 : cephadm [INF] Deploying daemon osd.1 on vm07 2026-03-07T07:55:28.636 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:28 vm07 bash[17031]: cluster 2026-03-07T06:55:27.083846+0000 mgr.vm07.yrfcuj (mgr.14201) 77 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:28.636 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:28 vm07 bash[17031]: cluster 2026-03-07T06:55:27.083846+0000 mgr.vm07.yrfcuj (mgr.14201) 77 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:28.636 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:28 vm07 bash[17031]: cephadm 2026-03-07T06:55:27.187478+0000 mgr.vm07.yrfcuj (mgr.14201) 78 : cephadm [INF] Deploying daemon osd.0 on vm09 2026-03-07T07:55:28.636 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:28 vm07 bash[17031]: cephadm 2026-03-07T06:55:27.187478+0000 mgr.vm07.yrfcuj (mgr.14201) 78 : cephadm [INF] Deploying daemon osd.0 on vm09 2026-03-07T07:55:28.636 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:28 vm07 bash[17031]: audit 2026-03-07T06:55:28.249707+0000 mon.vm07 (mon.0) 434 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:28.636 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:28 vm07 bash[17031]: audit 2026-03-07T06:55:28.249707+0000 mon.vm07 (mon.0) 434 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:28.636 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:28 vm07 bash[17031]: audit 2026-03-07T06:55:28.268274+0000 mon.vm07 (mon.0) 435 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:28.637 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:28 vm07 bash[17031]: audit 2026-03-07T06:55:28.268274+0000 mon.vm07 (mon.0) 435 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:28.637 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:28 vm07 bash[17031]: audit 2026-03-07T06:55:28.269225+0000 mon.vm07 (mon.0) 436 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-07T07:55:28.637 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:28 vm07 bash[17031]: audit 2026-03-07T06:55:28.269225+0000 mon.vm07 (mon.0) 436 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-07T07:55:28.637 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:28 vm07 bash[17031]: audit 2026-03-07T06:55:28.269773+0000 mon.vm07 (mon.0) 437 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:28.637 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:28 vm07 bash[17031]: audit 2026-03-07T06:55:28.269773+0000 mon.vm07 (mon.0) 437 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:29.572 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:29 vm07 bash[17031]: cephadm 2026-03-07T06:55:28.270414+0000 mgr.vm07.yrfcuj (mgr.14201) 79 : cephadm [INF] Deploying daemon osd.2 on vm07 2026-03-07T07:55:29.573 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:29 vm07 bash[17031]: cephadm 2026-03-07T06:55:28.270414+0000 mgr.vm07.yrfcuj (mgr.14201) 79 : cephadm [INF] Deploying daemon osd.2 on vm07 2026-03-07T07:55:29.573 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:29 vm07 bash[17031]: audit 2026-03-07T06:55:28.492059+0000 mon.vm07 (mon.0) 438 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:29.573 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:29 vm07 bash[17031]: audit 2026-03-07T06:55:28.492059+0000 mon.vm07 (mon.0) 438 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:29.573 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:29 vm07 bash[17031]: audit 2026-03-07T06:55:28.501328+0000 mon.vm07 (mon.0) 439 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:29.573 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:29 vm07 bash[17031]: audit 2026-03-07T06:55:28.501328+0000 mon.vm07 (mon.0) 439 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:29.573 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:29 vm07 bash[17031]: audit 2026-03-07T06:55:28.506966+0000 mon.vm07 (mon.0) 440 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-07T07:55:29.573 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:29 vm07 bash[17031]: audit 2026-03-07T06:55:28.506966+0000 mon.vm07 (mon.0) 440 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-07T07:55:29.573 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:29 vm07 bash[17031]: audit 2026-03-07T06:55:28.507628+0000 mon.vm07 (mon.0) 441 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:29.573 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:29 vm07 bash[17031]: audit 2026-03-07T06:55:28.507628+0000 mon.vm07 (mon.0) 441 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:29.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:29 vm09 bash[21220]: cephadm 2026-03-07T06:55:28.270414+0000 mgr.vm07.yrfcuj (mgr.14201) 79 : cephadm [INF] Deploying daemon osd.2 on vm07 2026-03-07T07:55:29.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:29 vm09 bash[21220]: cephadm 2026-03-07T06:55:28.270414+0000 mgr.vm07.yrfcuj (mgr.14201) 79 : cephadm [INF] Deploying daemon osd.2 on vm07 2026-03-07T07:55:29.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:29 vm09 bash[21220]: audit 2026-03-07T06:55:28.492059+0000 mon.vm07 (mon.0) 438 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:29.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:29 vm09 bash[21220]: audit 2026-03-07T06:55:28.492059+0000 mon.vm07 (mon.0) 438 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:29.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:29 vm09 bash[21220]: audit 2026-03-07T06:55:28.501328+0000 mon.vm07 (mon.0) 439 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:29.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:29 vm09 bash[21220]: audit 2026-03-07T06:55:28.501328+0000 mon.vm07 (mon.0) 439 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:29.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:29 vm09 bash[21220]: audit 2026-03-07T06:55:28.506966+0000 mon.vm07 (mon.0) 440 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-07T07:55:29.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:29 vm09 bash[21220]: audit 2026-03-07T06:55:28.506966+0000 mon.vm07 (mon.0) 440 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-07T07:55:29.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:29 vm09 bash[21220]: audit 2026-03-07T06:55:28.507628+0000 mon.vm07 (mon.0) 441 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:29.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:29 vm09 bash[21220]: audit 2026-03-07T06:55:28.507628+0000 mon.vm07 (mon.0) 441 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:29.830 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:29 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:55:30.063 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:29 vm09 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:55:30.092 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:29 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:55:30.243 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:55:30.316 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:55:30.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: cephadm 2026-03-07T06:55:28.508075+0000 mgr.vm07.yrfcuj (mgr.14201) 80 : cephadm [INF] Deploying daemon osd.3 on vm09 2026-03-07T07:55:30.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: cephadm 2026-03-07T06:55:28.508075+0000 mgr.vm07.yrfcuj (mgr.14201) 80 : cephadm [INF] Deploying daemon osd.3 on vm09 2026-03-07T07:55:30.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: cluster 2026-03-07T06:55:29.084068+0000 mgr.vm07.yrfcuj (mgr.14201) 81 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:30.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: cluster 2026-03-07T06:55:29.084068+0000 mgr.vm07.yrfcuj (mgr.14201) 81 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:30.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: audit 2026-03-07T06:55:29.950663+0000 mon.vm07 (mon.0) 442 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:30.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: audit 2026-03-07T06:55:29.950663+0000 mon.vm07 (mon.0) 442 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:30.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: audit 2026-03-07T06:55:29.957366+0000 mon.vm07 (mon.0) 443 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:30.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: audit 2026-03-07T06:55:29.957366+0000 mon.vm07 (mon.0) 443 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:30.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: audit 2026-03-07T06:55:29.958016+0000 mon.vm07 (mon.0) 444 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-07T07:55:30.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: audit 2026-03-07T06:55:29.958016+0000 mon.vm07 (mon.0) 444 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-07T07:55:30.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: audit 2026-03-07T06:55:29.958597+0000 mon.vm07 (mon.0) 445 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:30.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: audit 2026-03-07T06:55:29.958597+0000 mon.vm07 (mon.0) 445 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:30.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: audit 2026-03-07T06:55:30.188691+0000 mon.vm07 (mon.0) 446 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:30.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: audit 2026-03-07T06:55:30.188691+0000 mon.vm07 (mon.0) 446 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:30.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: audit 2026-03-07T06:55:30.200250+0000 mon.vm07 (mon.0) 447 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:30.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: audit 2026-03-07T06:55:30.200250+0000 mon.vm07 (mon.0) 447 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:30.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: audit 2026-03-07T06:55:30.207557+0000 mon.vm07 (mon.0) 448 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-07T07:55:30.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: audit 2026-03-07T06:55:30.207557+0000 mon.vm07 (mon.0) 448 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-07T07:55:30.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: audit 2026-03-07T06:55:30.208078+0000 mon.vm07 (mon.0) 449 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:30.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:30 vm07 bash[17031]: audit 2026-03-07T06:55:30.208078+0000 mon.vm07 (mon.0) 449 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: cephadm 2026-03-07T06:55:28.508075+0000 mgr.vm07.yrfcuj (mgr.14201) 80 : cephadm [INF] Deploying daemon osd.3 on vm09 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: cephadm 2026-03-07T06:55:28.508075+0000 mgr.vm07.yrfcuj (mgr.14201) 80 : cephadm [INF] Deploying daemon osd.3 on vm09 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: cluster 2026-03-07T06:55:29.084068+0000 mgr.vm07.yrfcuj (mgr.14201) 81 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: cluster 2026-03-07T06:55:29.084068+0000 mgr.vm07.yrfcuj (mgr.14201) 81 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: audit 2026-03-07T06:55:29.950663+0000 mon.vm07 (mon.0) 442 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: audit 2026-03-07T06:55:29.950663+0000 mon.vm07 (mon.0) 442 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: audit 2026-03-07T06:55:29.957366+0000 mon.vm07 (mon.0) 443 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: audit 2026-03-07T06:55:29.957366+0000 mon.vm07 (mon.0) 443 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: audit 2026-03-07T06:55:29.958016+0000 mon.vm07 (mon.0) 444 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: audit 2026-03-07T06:55:29.958016+0000 mon.vm07 (mon.0) 444 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: audit 2026-03-07T06:55:29.958597+0000 mon.vm07 (mon.0) 445 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: audit 2026-03-07T06:55:29.958597+0000 mon.vm07 (mon.0) 445 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: audit 2026-03-07T06:55:30.188691+0000 mon.vm07 (mon.0) 446 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: audit 2026-03-07T06:55:30.188691+0000 mon.vm07 (mon.0) 446 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: audit 2026-03-07T06:55:30.200250+0000 mon.vm07 (mon.0) 447 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: audit 2026-03-07T06:55:30.200250+0000 mon.vm07 (mon.0) 447 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: audit 2026-03-07T06:55:30.207557+0000 mon.vm07 (mon.0) 448 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: audit 2026-03-07T06:55:30.207557+0000 mon.vm07 (mon.0) 448 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: audit 2026-03-07T06:55:30.208078+0000 mon.vm07 (mon.0) 449 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:30.799 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:30 vm09 bash[21220]: audit 2026-03-07T06:55:30.208078+0000 mon.vm07 (mon.0) 449 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:30.960 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:55:31.114 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":13,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1772866518,"num_remapped_pgs":0} 2026-03-07T07:55:31.526 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:31 vm07 bash[17031]: cephadm 2026-03-07T06:55:29.959331+0000 mgr.vm07.yrfcuj (mgr.14201) 82 : cephadm [INF] Deploying daemon osd.5 on vm07 2026-03-07T07:55:31.526 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:31 vm07 bash[17031]: cephadm 2026-03-07T06:55:29.959331+0000 mgr.vm07.yrfcuj (mgr.14201) 82 : cephadm [INF] Deploying daemon osd.5 on vm07 2026-03-07T07:55:31.526 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:31 vm07 bash[17031]: cephadm 2026-03-07T06:55:30.208478+0000 mgr.vm07.yrfcuj (mgr.14201) 83 : cephadm [INF] Deploying daemon osd.4 on vm09 2026-03-07T07:55:31.526 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:31 vm07 bash[17031]: cephadm 2026-03-07T06:55:30.208478+0000 mgr.vm07.yrfcuj (mgr.14201) 83 : cephadm [INF] Deploying daemon osd.4 on vm09 2026-03-07T07:55:31.526 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:31 vm07 bash[17031]: audit 2026-03-07T06:55:30.955254+0000 mon.vm07 (mon.0) 450 : audit [DBG] from='client.? 192.168.123.107:0/2904327659' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:31.526 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:31 vm07 bash[17031]: audit 2026-03-07T06:55:30.955254+0000 mon.vm07 (mon.0) 450 : audit [DBG] from='client.? 192.168.123.107:0/2904327659' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:31.822 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:31 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:55:31.875 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:31 vm09 bash[21220]: cephadm 2026-03-07T06:55:29.959331+0000 mgr.vm07.yrfcuj (mgr.14201) 82 : cephadm [INF] Deploying daemon osd.5 on vm07 2026-03-07T07:55:31.875 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:31 vm09 bash[21220]: cephadm 2026-03-07T06:55:29.959331+0000 mgr.vm07.yrfcuj (mgr.14201) 82 : cephadm [INF] Deploying daemon osd.5 on vm07 2026-03-07T07:55:31.875 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:31 vm09 bash[21220]: cephadm 2026-03-07T06:55:30.208478+0000 mgr.vm07.yrfcuj (mgr.14201) 83 : cephadm [INF] Deploying daemon osd.4 on vm09 2026-03-07T07:55:31.875 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:31 vm09 bash[21220]: cephadm 2026-03-07T06:55:30.208478+0000 mgr.vm07.yrfcuj (mgr.14201) 83 : cephadm [INF] Deploying daemon osd.4 on vm09 2026-03-07T07:55:31.875 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:31 vm09 bash[21220]: audit 2026-03-07T06:55:30.955254+0000 mon.vm07 (mon.0) 450 : audit [DBG] from='client.? 192.168.123.107:0/2904327659' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:31.875 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:31 vm09 bash[21220]: audit 2026-03-07T06:55:30.955254+0000 mon.vm07 (mon.0) 450 : audit [DBG] from='client.? 192.168.123.107:0/2904327659' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:32.073 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:31 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:55:32.114 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd stat -f json 2026-03-07T07:55:32.155 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:31 vm09 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:55:32.451 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: cluster 2026-03-07T06:55:31.084249+0000 mgr.vm07.yrfcuj (mgr.14201) 84 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: cluster 2026-03-07T06:55:31.084249+0000 mgr.vm07.yrfcuj (mgr.14201) 84 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: audit 2026-03-07T06:55:32.121430+0000 mon.vm07 (mon.0) 451 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: audit 2026-03-07T06:55:32.121430+0000 mon.vm07 (mon.0) 451 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: audit 2026-03-07T06:55:32.138693+0000 mon.vm07 (mon.0) 452 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: audit 2026-03-07T06:55:32.138693+0000 mon.vm07 (mon.0) 452 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: audit 2026-03-07T06:55:32.149312+0000 mon.vm07 (mon.0) 453 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: audit 2026-03-07T06:55:32.149312+0000 mon.vm07 (mon.0) 453 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: audit 2026-03-07T06:55:32.151897+0000 mon.vm07 (mon.0) 454 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: audit 2026-03-07T06:55:32.151897+0000 mon.vm07 (mon.0) 454 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: cephadm 2026-03-07T06:55:32.159130+0000 mgr.vm07.yrfcuj (mgr.14201) 85 : cephadm [INF] Deploying daemon osd.6 on vm07 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: cephadm 2026-03-07T06:55:32.159130+0000 mgr.vm07.yrfcuj (mgr.14201) 85 : cephadm [INF] Deploying daemon osd.6 on vm07 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: audit 2026-03-07T06:55:32.305044+0000 mon.vm07 (mon.0) 455 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: audit 2026-03-07T06:55:32.305044+0000 mon.vm07 (mon.0) 455 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: audit 2026-03-07T06:55:32.320897+0000 mon.vm07 (mon.0) 456 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: audit 2026-03-07T06:55:32.320897+0000 mon.vm07 (mon.0) 456 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: audit 2026-03-07T06:55:32.323162+0000 mon.vm07 (mon.0) 457 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: audit 2026-03-07T06:55:32.323162+0000 mon.vm07 (mon.0) 457 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: audit 2026-03-07T06:55:32.340104+0000 mon.vm07 (mon.0) 458 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: audit 2026-03-07T06:55:32.340104+0000 mon.vm07 (mon.0) 458 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: cephadm 2026-03-07T06:55:32.355433+0000 mgr.vm07.yrfcuj (mgr.14201) 86 : cephadm [INF] Deploying daemon osd.7 on vm09 2026-03-07T07:55:32.660 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:32 vm07 bash[17031]: cephadm 2026-03-07T06:55:32.355433+0000 mgr.vm07.yrfcuj (mgr.14201) 86 : cephadm [INF] Deploying daemon osd.7 on vm09 2026-03-07T07:55:32.818 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: cluster 2026-03-07T06:55:31.084249+0000 mgr.vm07.yrfcuj (mgr.14201) 84 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:32.818 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: cluster 2026-03-07T06:55:31.084249+0000 mgr.vm07.yrfcuj (mgr.14201) 84 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:32.818 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: audit 2026-03-07T06:55:32.121430+0000 mon.vm07 (mon.0) 451 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:32.818 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: audit 2026-03-07T06:55:32.121430+0000 mon.vm07 (mon.0) 451 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:32.818 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: audit 2026-03-07T06:55:32.138693+0000 mon.vm07 (mon.0) 452 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:32.818 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: audit 2026-03-07T06:55:32.138693+0000 mon.vm07 (mon.0) 452 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:32.818 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: audit 2026-03-07T06:55:32.149312+0000 mon.vm07 (mon.0) 453 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-07T07:55:32.818 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: audit 2026-03-07T06:55:32.149312+0000 mon.vm07 (mon.0) 453 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-07T07:55:32.818 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: audit 2026-03-07T06:55:32.151897+0000 mon.vm07 (mon.0) 454 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:32.818 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: audit 2026-03-07T06:55:32.151897+0000 mon.vm07 (mon.0) 454 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:32.818 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: cephadm 2026-03-07T06:55:32.159130+0000 mgr.vm07.yrfcuj (mgr.14201) 85 : cephadm [INF] Deploying daemon osd.6 on vm07 2026-03-07T07:55:32.818 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: cephadm 2026-03-07T06:55:32.159130+0000 mgr.vm07.yrfcuj (mgr.14201) 85 : cephadm [INF] Deploying daemon osd.6 on vm07 2026-03-07T07:55:32.818 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: audit 2026-03-07T06:55:32.305044+0000 mon.vm07 (mon.0) 455 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:32.818 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: audit 2026-03-07T06:55:32.305044+0000 mon.vm07 (mon.0) 455 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:32.819 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: audit 2026-03-07T06:55:32.320897+0000 mon.vm07 (mon.0) 456 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:32.819 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: audit 2026-03-07T06:55:32.320897+0000 mon.vm07 (mon.0) 456 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:32.819 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: audit 2026-03-07T06:55:32.323162+0000 mon.vm07 (mon.0) 457 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-07T07:55:32.819 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: audit 2026-03-07T06:55:32.323162+0000 mon.vm07 (mon.0) 457 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-07T07:55:32.819 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: audit 2026-03-07T06:55:32.340104+0000 mon.vm07 (mon.0) 458 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:32.819 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: audit 2026-03-07T06:55:32.340104+0000 mon.vm07 (mon.0) 458 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:32.819 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: cephadm 2026-03-07T06:55:32.355433+0000 mgr.vm07.yrfcuj (mgr.14201) 86 : cephadm [INF] Deploying daemon osd.7 on vm09 2026-03-07T07:55:32.819 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:32 vm09 bash[21220]: cephadm 2026-03-07T06:55:32.355433+0000 mgr.vm07.yrfcuj (mgr.14201) 86 : cephadm [INF] Deploying daemon osd.7 on vm09 2026-03-07T07:55:33.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:33 vm07 bash[17031]: cluster 2026-03-07T06:55:33.084882+0000 mgr.vm07.yrfcuj (mgr.14201) 87 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:33.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:33 vm07 bash[17031]: cluster 2026-03-07T06:55:33.084882+0000 mgr.vm07.yrfcuj (mgr.14201) 87 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:33.756 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:33 vm09 bash[21220]: cluster 2026-03-07T06:55:33.084882+0000 mgr.vm07.yrfcuj (mgr.14201) 87 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:33.756 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:33 vm09 bash[21220]: cluster 2026-03-07T06:55:33.084882+0000 mgr.vm07.yrfcuj (mgr.14201) 87 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:34.044 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:33 vm09 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:55:34.265 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:34 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:55:34.332 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:34 vm09 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:55:34.547 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:34 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:55:34.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:34 vm09 bash[21220]: audit 2026-03-07T06:55:34.148659+0000 mon.vm07 (mon.0) 459 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-07T07:55:34.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:34 vm09 bash[21220]: audit 2026-03-07T06:55:34.148659+0000 mon.vm07 (mon.0) 459 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-07T07:55:34.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:34 vm09 bash[21220]: audit 2026-03-07T06:55:34.153206+0000 mon.vm09 (mon.1) 10 : audit [INF] from='osd.0 [v2:192.168.123.109:6800/635544560,v1:192.168.123.109:6801/635544560]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-07T07:55:34.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:34 vm09 bash[21220]: audit 2026-03-07T06:55:34.153206+0000 mon.vm09 (mon.1) 10 : audit [INF] from='osd.0 [v2:192.168.123.109:6800/635544560,v1:192.168.123.109:6801/635544560]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-07T07:55:34.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:34 vm09 bash[21220]: audit 2026-03-07T06:55:34.283483+0000 mon.vm07 (mon.0) 460 : audit [INF] from='osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-07T07:55:34.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:34 vm09 bash[21220]: audit 2026-03-07T06:55:34.283483+0000 mon.vm07 (mon.0) 460 : audit [INF] from='osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-07T07:55:34.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:34 vm09 bash[21220]: audit 2026-03-07T06:55:34.372268+0000 mon.vm07 (mon.0) 461 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:34.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:34 vm09 bash[21220]: audit 2026-03-07T06:55:34.372268+0000 mon.vm07 (mon.0) 461 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:34.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:34 vm09 bash[21220]: audit 2026-03-07T06:55:34.387231+0000 mon.vm07 (mon.0) 462 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:34.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:34 vm09 bash[21220]: audit 2026-03-07T06:55:34.387231+0000 mon.vm07 (mon.0) 462 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:34.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:34 vm09 bash[21220]: audit 2026-03-07T06:55:34.461073+0000 mon.vm07 (mon.0) 463 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:34 vm09 bash[21220]: audit 2026-03-07T06:55:34.461073+0000 mon.vm07 (mon.0) 463 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:34 vm09 bash[21220]: audit 2026-03-07T06:55:34.495374+0000 mon.vm07 (mon.0) 464 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:34 vm09 bash[21220]: audit 2026-03-07T06:55:34.495374+0000 mon.vm07 (mon.0) 464 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:34.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:34 vm07 bash[17031]: audit 2026-03-07T06:55:34.148659+0000 mon.vm07 (mon.0) 459 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-07T07:55:34.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:34 vm07 bash[17031]: audit 2026-03-07T06:55:34.148659+0000 mon.vm07 (mon.0) 459 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-07T07:55:34.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:34 vm07 bash[17031]: audit 2026-03-07T06:55:34.153206+0000 mon.vm09 (mon.1) 10 : audit [INF] from='osd.0 [v2:192.168.123.109:6800/635544560,v1:192.168.123.109:6801/635544560]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-07T07:55:34.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:34 vm07 bash[17031]: audit 2026-03-07T06:55:34.153206+0000 mon.vm09 (mon.1) 10 : audit [INF] from='osd.0 [v2:192.168.123.109:6800/635544560,v1:192.168.123.109:6801/635544560]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-07T07:55:34.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:34 vm07 bash[17031]: audit 2026-03-07T06:55:34.283483+0000 mon.vm07 (mon.0) 460 : audit [INF] from='osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-07T07:55:34.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:34 vm07 bash[17031]: audit 2026-03-07T06:55:34.283483+0000 mon.vm07 (mon.0) 460 : audit [INF] from='osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-07T07:55:34.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:34 vm07 bash[17031]: audit 2026-03-07T06:55:34.372268+0000 mon.vm07 (mon.0) 461 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:34.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:34 vm07 bash[17031]: audit 2026-03-07T06:55:34.372268+0000 mon.vm07 (mon.0) 461 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:34.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:34 vm07 bash[17031]: audit 2026-03-07T06:55:34.387231+0000 mon.vm07 (mon.0) 462 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:34.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:34 vm07 bash[17031]: audit 2026-03-07T06:55:34.387231+0000 mon.vm07 (mon.0) 462 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:34.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:34 vm07 bash[17031]: audit 2026-03-07T06:55:34.461073+0000 mon.vm07 (mon.0) 463 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:34.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:34 vm07 bash[17031]: audit 2026-03-07T06:55:34.461073+0000 mon.vm07 (mon.0) 463 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:34.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:34 vm07 bash[17031]: audit 2026-03-07T06:55:34.495374+0000 mon.vm07 (mon.0) 464 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:34.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:34 vm07 bash[17031]: audit 2026-03-07T06:55:34.495374+0000 mon.vm07 (mon.0) 464 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.557771+0000 mon.vm07 (mon.0) 465 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.557771+0000 mon.vm07 (mon.0) 465 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.557843+0000 mon.vm07 (mon.0) 466 : audit [INF] from='osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.557843+0000 mon.vm07 (mon.0) 466 : audit [INF] from='osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: cluster 2026-03-07T06:55:34.560589+0000 mon.vm07 (mon.0) 467 : cluster [DBG] osdmap e14: 8 total, 0 up, 8 in 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: cluster 2026-03-07T06:55:34.560589+0000 mon.vm07 (mon.0) 467 : cluster [DBG] osdmap e14: 8 total, 0 up, 8 in 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.560911+0000 mon.vm07 (mon.0) 468 : audit [INF] from='osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.560911+0000 mon.vm07 (mon.0) 468 : audit [INF] from='osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.570005+0000 mon.vm09 (mon.1) 11 : audit [INF] from='osd.0 [v2:192.168.123.109:6800/635544560,v1:192.168.123.109:6801/635544560]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.570005+0000 mon.vm09 (mon.1) 11 : audit [INF] from='osd.0 [v2:192.168.123.109:6800/635544560,v1:192.168.123.109:6801/635544560]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.571299+0000 mon.vm07 (mon.0) 469 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.571299+0000 mon.vm07 (mon.0) 469 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.571447+0000 mon.vm07 (mon.0) 470 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.571447+0000 mon.vm07 (mon.0) 470 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.571531+0000 mon.vm07 (mon.0) 471 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.571531+0000 mon.vm07 (mon.0) 471 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.571613+0000 mon.vm07 (mon.0) 472 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.571613+0000 mon.vm07 (mon.0) 472 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.571646+0000 mon.vm07 (mon.0) 473 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.571646+0000 mon.vm07 (mon.0) 473 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.571673+0000 mon.vm07 (mon.0) 474 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.571673+0000 mon.vm07 (mon.0) 474 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.571701+0000 mon.vm07 (mon.0) 475 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.571701+0000 mon.vm07 (mon.0) 475 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.571727+0000 mon.vm07 (mon.0) 476 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.571727+0000 mon.vm07 (mon.0) 476 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.571752+0000 mon.vm07 (mon.0) 477 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: audit 2026-03-07T06:55:34.571752+0000 mon.vm07 (mon.0) 477 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:35.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: cluster 2026-03-07T06:55:35.085119+0000 mgr.vm07.yrfcuj (mgr.14201) 88 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:35.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:35 vm07 bash[17031]: cluster 2026-03-07T06:55:35.085119+0000 mgr.vm07.yrfcuj (mgr.14201) 88 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:35.734 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.557771+0000 mon.vm07 (mon.0) 465 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.557771+0000 mon.vm07 (mon.0) 465 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.557843+0000 mon.vm07 (mon.0) 466 : audit [INF] from='osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.557843+0000 mon.vm07 (mon.0) 466 : audit [INF] from='osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: cluster 2026-03-07T06:55:34.560589+0000 mon.vm07 (mon.0) 467 : cluster [DBG] osdmap e14: 8 total, 0 up, 8 in 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: cluster 2026-03-07T06:55:34.560589+0000 mon.vm07 (mon.0) 467 : cluster [DBG] osdmap e14: 8 total, 0 up, 8 in 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.560911+0000 mon.vm07 (mon.0) 468 : audit [INF] from='osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.560911+0000 mon.vm07 (mon.0) 468 : audit [INF] from='osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.570005+0000 mon.vm09 (mon.1) 11 : audit [INF] from='osd.0 [v2:192.168.123.109:6800/635544560,v1:192.168.123.109:6801/635544560]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.570005+0000 mon.vm09 (mon.1) 11 : audit [INF] from='osd.0 [v2:192.168.123.109:6800/635544560,v1:192.168.123.109:6801/635544560]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.571299+0000 mon.vm07 (mon.0) 469 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.571299+0000 mon.vm07 (mon.0) 469 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.571447+0000 mon.vm07 (mon.0) 470 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.571447+0000 mon.vm07 (mon.0) 470 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.571531+0000 mon.vm07 (mon.0) 471 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.571531+0000 mon.vm07 (mon.0) 471 : audit [INF] from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.571613+0000 mon.vm07 (mon.0) 472 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.571613+0000 mon.vm07 (mon.0) 472 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.571646+0000 mon.vm07 (mon.0) 473 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.571646+0000 mon.vm07 (mon.0) 473 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.571673+0000 mon.vm07 (mon.0) 474 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.571673+0000 mon.vm07 (mon.0) 474 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.571701+0000 mon.vm07 (mon.0) 475 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.571701+0000 mon.vm07 (mon.0) 475 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.571727+0000 mon.vm07 (mon.0) 476 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.571727+0000 mon.vm07 (mon.0) 476 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.571752+0000 mon.vm07 (mon.0) 477 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: audit 2026-03-07T06:55:34.571752+0000 mon.vm07 (mon.0) 477 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: cluster 2026-03-07T06:55:35.085119+0000 mgr.vm07.yrfcuj (mgr.14201) 88 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:35.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:35 vm09 bash[21220]: cluster 2026-03-07T06:55:35.085119+0000 mgr.vm07.yrfcuj (mgr.14201) 88 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:36.815 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.561655+0000 mon.vm07 (mon.0) 478 : audit [INF] from='osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-07T07:55:36.815 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.561655+0000 mon.vm07 (mon.0) 478 : audit [INF] from='osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-07T07:55:36.815 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.561729+0000 mon.vm07 (mon.0) 479 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-07T07:55:36.815 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.561729+0000 mon.vm07 (mon.0) 479 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-07T07:55:36.815 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: cluster 2026-03-07T06:55:35.564063+0000 mon.vm07 (mon.0) 480 : cluster [DBG] osdmap e15: 8 total, 0 up, 8 in 2026-03-07T07:55:36.815 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: cluster 2026-03-07T06:55:35.564063+0000 mon.vm07 (mon.0) 480 : cluster [DBG] osdmap e15: 8 total, 0 up, 8 in 2026-03-07T07:55:36.815 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.564951+0000 mon.vm07 (mon.0) 481 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:36.815 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.564951+0000 mon.vm07 (mon.0) 481 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:36.815 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.565044+0000 mon.vm07 (mon.0) 482 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:36.815 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.565044+0000 mon.vm07 (mon.0) 482 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.565105+0000 mon.vm07 (mon.0) 483 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.565105+0000 mon.vm07 (mon.0) 483 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.565141+0000 mon.vm07 (mon.0) 484 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.565141+0000 mon.vm07 (mon.0) 484 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.565170+0000 mon.vm07 (mon.0) 485 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.565170+0000 mon.vm07 (mon.0) 485 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.565196+0000 mon.vm07 (mon.0) 486 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.565196+0000 mon.vm07 (mon.0) 486 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.565223+0000 mon.vm07 (mon.0) 487 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.565223+0000 mon.vm07 (mon.0) 487 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.565249+0000 mon.vm07 (mon.0) 488 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.565249+0000 mon.vm07 (mon.0) 488 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.567599+0000 mon.vm07 (mon.0) 489 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.567599+0000 mon.vm07 (mon.0) 489 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.580652+0000 mon.vm07 (mon.0) 490 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.580652+0000 mon.vm07 (mon.0) 490 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.739513+0000 mon.vm07 (mon.0) 491 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:35.739513+0000 mon.vm07 (mon.0) 491 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:36.567243+0000 mon.vm07 (mon.0) 492 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:36.816 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:36 vm09 bash[21220]: audit 2026-03-07T06:55:36.567243+0000 mon.vm07 (mon.0) 492 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.561655+0000 mon.vm07 (mon.0) 478 : audit [INF] from='osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.561655+0000 mon.vm07 (mon.0) 478 : audit [INF] from='osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.561729+0000 mon.vm07 (mon.0) 479 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.561729+0000 mon.vm07 (mon.0) 479 : audit [INF] from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: cluster 2026-03-07T06:55:35.564063+0000 mon.vm07 (mon.0) 480 : cluster [DBG] osdmap e15: 8 total, 0 up, 8 in 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: cluster 2026-03-07T06:55:35.564063+0000 mon.vm07 (mon.0) 480 : cluster [DBG] osdmap e15: 8 total, 0 up, 8 in 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.564951+0000 mon.vm07 (mon.0) 481 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.564951+0000 mon.vm07 (mon.0) 481 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.565044+0000 mon.vm07 (mon.0) 482 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.565044+0000 mon.vm07 (mon.0) 482 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.565105+0000 mon.vm07 (mon.0) 483 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.565105+0000 mon.vm07 (mon.0) 483 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.565141+0000 mon.vm07 (mon.0) 484 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.565141+0000 mon.vm07 (mon.0) 484 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.565170+0000 mon.vm07 (mon.0) 485 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.565170+0000 mon.vm07 (mon.0) 485 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.565196+0000 mon.vm07 (mon.0) 486 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.565196+0000 mon.vm07 (mon.0) 486 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.565223+0000 mon.vm07 (mon.0) 487 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.565223+0000 mon.vm07 (mon.0) 487 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.565249+0000 mon.vm07 (mon.0) 488 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.565249+0000 mon.vm07 (mon.0) 488 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.567599+0000 mon.vm07 (mon.0) 489 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.567599+0000 mon.vm07 (mon.0) 489 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.580652+0000 mon.vm07 (mon.0) 490 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.580652+0000 mon.vm07 (mon.0) 490 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.739513+0000 mon.vm07 (mon.0) 491 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:35.739513+0000 mon.vm07 (mon.0) 491 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:36.567243+0000 mon.vm07 (mon.0) 492 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:36.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:36 vm07 bash[17031]: audit 2026-03-07T06:55:36.567243+0000 mon.vm07 (mon.0) 492 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: cluster 2026-03-07T06:55:35.175089+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: cluster 2026-03-07T06:55:35.175089+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: cluster 2026-03-07T06:55:35.175153+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: cluster 2026-03-07T06:55:35.175153+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: cluster 2026-03-07T06:55:35.231096+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: cluster 2026-03-07T06:55:35.231096+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: cluster 2026-03-07T06:55:35.231150+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: cluster 2026-03-07T06:55:35.231150+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.574623+0000 mon.vm07 (mon.0) 493 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.574623+0000 mon.vm07 (mon.0) 493 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.581948+0000 mon.vm07 (mon.0) 494 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.581948+0000 mon.vm07 (mon.0) 494 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: cluster 2026-03-07T06:55:36.587229+0000 mon.vm07 (mon.0) 495 : cluster [INF] osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522] boot 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: cluster 2026-03-07T06:55:36.587229+0000 mon.vm07 (mon.0) 495 : cluster [INF] osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522] boot 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: cluster 2026-03-07T06:55:36.587256+0000 mon.vm07 (mon.0) 496 : cluster [DBG] osdmap e16: 8 total, 1 up, 8 in 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: cluster 2026-03-07T06:55:36.587256+0000 mon.vm07 (mon.0) 496 : cluster [DBG] osdmap e16: 8 total, 1 up, 8 in 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.594319+0000 mon.vm07 (mon.0) 497 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.594319+0000 mon.vm07 (mon.0) 497 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.594482+0000 mon.vm07 (mon.0) 498 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.594482+0000 mon.vm07 (mon.0) 498 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.594578+0000 mon.vm07 (mon.0) 499 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.594578+0000 mon.vm07 (mon.0) 499 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.594655+0000 mon.vm07 (mon.0) 500 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.594655+0000 mon.vm07 (mon.0) 500 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.594695+0000 mon.vm07 (mon.0) 501 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.594695+0000 mon.vm07 (mon.0) 501 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.594733+0000 mon.vm07 (mon.0) 502 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.594733+0000 mon.vm07 (mon.0) 502 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.594894+0000 mon.vm07 (mon.0) 503 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.594894+0000 mon.vm07 (mon.0) 503 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.594936+0000 mon.vm07 (mon.0) 504 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.594936+0000 mon.vm07 (mon.0) 504 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.594968+0000 mon.vm07 (mon.0) 505 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.594968+0000 mon.vm07 (mon.0) 505 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.637186+0000 mon.vm07 (mon.0) 506 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.637186+0000 mon.vm07 (mon.0) 506 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.641848+0000 mon.vm09 (mon.1) 12 : audit [INF] from='osd.2 [v2:192.168.123.107:6810/158678589,v1:192.168.123.107:6811/158678589]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.641848+0000 mon.vm09 (mon.1) 12 : audit [INF] from='osd.2 [v2:192.168.123.107:6810/158678589,v1:192.168.123.107:6811/158678589]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.814085+0000 mon.vm07 (mon.0) 507 : audit [INF] from='osd.0 ' entity='osd.0' 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:36.814085+0000 mon.vm07 (mon.0) 507 : audit [INF] from='osd.0 ' entity='osd.0' 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: cluster 2026-03-07T06:55:37.085313+0000 mgr.vm07.yrfcuj (mgr.14201) 89 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: cluster 2026-03-07T06:55:37.085313+0000 mgr.vm07.yrfcuj (mgr.14201) 89 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.573399+0000 mon.vm07 (mon.0) 508 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.573399+0000 mon.vm07 (mon.0) 508 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.585175+0000 mon.vm07 (mon.0) 509 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.585175+0000 mon.vm07 (mon.0) 509 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.585239+0000 mon.vm07 (mon.0) 510 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.585239+0000 mon.vm07 (mon.0) 510 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: cluster 2026-03-07T06:55:37.587304+0000 mon.vm07 (mon.0) 511 : cluster [INF] osd.0 [v2:192.168.123.109:6800/635544560,v1:192.168.123.109:6801/635544560] boot 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: cluster 2026-03-07T06:55:37.587304+0000 mon.vm07 (mon.0) 511 : cluster [INF] osd.0 [v2:192.168.123.109:6800/635544560,v1:192.168.123.109:6801/635544560] boot 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: cluster 2026-03-07T06:55:37.587351+0000 mon.vm07 (mon.0) 512 : cluster [DBG] osdmap e17: 8 total, 2 up, 8 in 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: cluster 2026-03-07T06:55:37.587351+0000 mon.vm07 (mon.0) 512 : cluster [DBG] osdmap e17: 8 total, 2 up, 8 in 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.587467+0000 mon.vm07 (mon.0) 513 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.587467+0000 mon.vm07 (mon.0) 513 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.587786+0000 mon.vm07 (mon.0) 514 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.587786+0000 mon.vm07 (mon.0) 514 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.587832+0000 mon.vm07 (mon.0) 515 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.587832+0000 mon.vm07 (mon.0) 515 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.587859+0000 mon.vm07 (mon.0) 516 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.587859+0000 mon.vm07 (mon.0) 516 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.587886+0000 mon.vm07 (mon.0) 517 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.587886+0000 mon.vm07 (mon.0) 517 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.587912+0000 mon.vm07 (mon.0) 518 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.587912+0000 mon.vm07 (mon.0) 518 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.587940+0000 mon.vm07 (mon.0) 519 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.587940+0000 mon.vm07 (mon.0) 519 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.589536+0000 mon.vm07 (mon.0) 520 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.589536+0000 mon.vm07 (mon.0) 520 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.592661+0000 mon.vm07 (mon.0) 521 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.592661+0000 mon.vm07 (mon.0) 521 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.594327+0000 mon.vm09 (mon.1) 13 : audit [INF] from='osd.2 [v2:192.168.123.107:6810/158678589,v1:192.168.123.107:6811/158678589]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:37.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:37 vm09 bash[21220]: audit 2026-03-07T06:55:37.594327+0000 mon.vm09 (mon.1) 13 : audit [INF] from='osd.2 [v2:192.168.123.107:6810/158678589,v1:192.168.123.107:6811/158678589]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:37.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: cluster 2026-03-07T06:55:35.175089+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: cluster 2026-03-07T06:55:35.175089+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: cluster 2026-03-07T06:55:35.175153+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: cluster 2026-03-07T06:55:35.175153+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: cluster 2026-03-07T06:55:35.231096+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: cluster 2026-03-07T06:55:35.231096+0000 osd.1 (osd.1) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: cluster 2026-03-07T06:55:35.231150+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: cluster 2026-03-07T06:55:35.231150+0000 osd.1 (osd.1) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.574623+0000 mon.vm07 (mon.0) 493 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.574623+0000 mon.vm07 (mon.0) 493 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.581948+0000 mon.vm07 (mon.0) 494 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.581948+0000 mon.vm07 (mon.0) 494 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: cluster 2026-03-07T06:55:36.587229+0000 mon.vm07 (mon.0) 495 : cluster [INF] osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522] boot 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: cluster 2026-03-07T06:55:36.587229+0000 mon.vm07 (mon.0) 495 : cluster [INF] osd.1 [v2:192.168.123.107:6802/1359088522,v1:192.168.123.107:6803/1359088522] boot 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: cluster 2026-03-07T06:55:36.587256+0000 mon.vm07 (mon.0) 496 : cluster [DBG] osdmap e16: 8 total, 1 up, 8 in 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: cluster 2026-03-07T06:55:36.587256+0000 mon.vm07 (mon.0) 496 : cluster [DBG] osdmap e16: 8 total, 1 up, 8 in 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.594319+0000 mon.vm07 (mon.0) 497 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.594319+0000 mon.vm07 (mon.0) 497 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.594482+0000 mon.vm07 (mon.0) 498 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.594482+0000 mon.vm07 (mon.0) 498 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.594578+0000 mon.vm07 (mon.0) 499 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.594578+0000 mon.vm07 (mon.0) 499 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.594655+0000 mon.vm07 (mon.0) 500 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.594655+0000 mon.vm07 (mon.0) 500 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.594695+0000 mon.vm07 (mon.0) 501 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.594695+0000 mon.vm07 (mon.0) 501 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.594733+0000 mon.vm07 (mon.0) 502 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.594733+0000 mon.vm07 (mon.0) 502 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.594894+0000 mon.vm07 (mon.0) 503 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.594894+0000 mon.vm07 (mon.0) 503 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.594936+0000 mon.vm07 (mon.0) 504 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.594936+0000 mon.vm07 (mon.0) 504 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.594968+0000 mon.vm07 (mon.0) 505 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.594968+0000 mon.vm07 (mon.0) 505 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.637186+0000 mon.vm07 (mon.0) 506 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.637186+0000 mon.vm07 (mon.0) 506 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.641848+0000 mon.vm09 (mon.1) 12 : audit [INF] from='osd.2 [v2:192.168.123.107:6810/158678589,v1:192.168.123.107:6811/158678589]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.641848+0000 mon.vm09 (mon.1) 12 : audit [INF] from='osd.2 [v2:192.168.123.107:6810/158678589,v1:192.168.123.107:6811/158678589]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.814085+0000 mon.vm07 (mon.0) 507 : audit [INF] from='osd.0 ' entity='osd.0' 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:36.814085+0000 mon.vm07 (mon.0) 507 : audit [INF] from='osd.0 ' entity='osd.0' 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: cluster 2026-03-07T06:55:37.085313+0000 mgr.vm07.yrfcuj (mgr.14201) 89 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: cluster 2026-03-07T06:55:37.085313+0000 mgr.vm07.yrfcuj (mgr.14201) 89 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.573399+0000 mon.vm07 (mon.0) 508 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.573399+0000 mon.vm07 (mon.0) 508 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.585175+0000 mon.vm07 (mon.0) 509 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.585175+0000 mon.vm07 (mon.0) 509 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.585239+0000 mon.vm07 (mon.0) 510 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.585239+0000 mon.vm07 (mon.0) 510 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: cluster 2026-03-07T06:55:37.587304+0000 mon.vm07 (mon.0) 511 : cluster [INF] osd.0 [v2:192.168.123.109:6800/635544560,v1:192.168.123.109:6801/635544560] boot 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: cluster 2026-03-07T06:55:37.587304+0000 mon.vm07 (mon.0) 511 : cluster [INF] osd.0 [v2:192.168.123.109:6800/635544560,v1:192.168.123.109:6801/635544560] boot 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: cluster 2026-03-07T06:55:37.587351+0000 mon.vm07 (mon.0) 512 : cluster [DBG] osdmap e17: 8 total, 2 up, 8 in 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: cluster 2026-03-07T06:55:37.587351+0000 mon.vm07 (mon.0) 512 : cluster [DBG] osdmap e17: 8 total, 2 up, 8 in 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.587467+0000 mon.vm07 (mon.0) 513 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:37.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.587467+0000 mon.vm07 (mon.0) 513 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-07T07:55:37.910 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.587786+0000 mon.vm07 (mon.0) 514 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:37.910 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.587786+0000 mon.vm07 (mon.0) 514 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:37.910 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.587832+0000 mon.vm07 (mon.0) 515 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:37.910 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.587832+0000 mon.vm07 (mon.0) 515 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:37.910 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.587859+0000 mon.vm07 (mon.0) 516 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:37.910 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.587859+0000 mon.vm07 (mon.0) 516 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:37.910 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.587886+0000 mon.vm07 (mon.0) 517 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:37.910 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.587886+0000 mon.vm07 (mon.0) 517 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:37.910 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.587912+0000 mon.vm07 (mon.0) 518 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:37.910 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.587912+0000 mon.vm07 (mon.0) 518 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:37.910 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.587940+0000 mon.vm07 (mon.0) 519 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:37.910 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.587940+0000 mon.vm07 (mon.0) 519 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:37.910 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.589536+0000 mon.vm07 (mon.0) 520 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:37.910 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.589536+0000 mon.vm07 (mon.0) 520 : audit [INF] from='osd.2 ' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:37.910 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.592661+0000 mon.vm07 (mon.0) 521 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:37.910 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.592661+0000 mon.vm07 (mon.0) 521 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:37.910 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.594327+0000 mon.vm09 (mon.1) 13 : audit [INF] from='osd.2 [v2:192.168.123.107:6810/158678589,v1:192.168.123.107:6811/158678589]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:37.910 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:37 vm07 bash[17031]: audit 2026-03-07T06:55:37.594327+0000 mon.vm09 (mon.1) 13 : audit [INF] from='osd.2 [v2:192.168.123.107:6810/158678589,v1:192.168.123.107:6811/158678589]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:37.748807+0000 mon.vm07 (mon.0) 522 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:37.748807+0000 mon.vm07 (mon.0) 522 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.400860+0000 mon.vm07 (mon.0) 523 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.400860+0000 mon.vm07 (mon.0) 523 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.574473+0000 mon.vm07 (mon.0) 524 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.574473+0000 mon.vm07 (mon.0) 524 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.588224+0000 mon.vm07 (mon.0) 525 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.588224+0000 mon.vm07 (mon.0) 525 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.588327+0000 mon.vm07 (mon.0) 526 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.588327+0000 mon.vm07 (mon.0) 526 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.588398+0000 mon.vm07 (mon.0) 527 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.588398+0000 mon.vm07 (mon.0) 527 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: cluster 2026-03-07T06:55:38.591389+0000 mon.vm07 (mon.0) 528 : cluster [DBG] osdmap e18: 8 total, 2 up, 8 in 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: cluster 2026-03-07T06:55:38.591389+0000 mon.vm07 (mon.0) 528 : cluster [DBG] osdmap e18: 8 total, 2 up, 8 in 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.591578+0000 mon.vm07 (mon.0) 529 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.591578+0000 mon.vm07 (mon.0) 529 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.591726+0000 mon.vm07 (mon.0) 530 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.591726+0000 mon.vm07 (mon.0) 530 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.591791+0000 mon.vm07 (mon.0) 531 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.591791+0000 mon.vm07 (mon.0) 531 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.591838+0000 mon.vm07 (mon.0) 532 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:38.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.591838+0000 mon.vm07 (mon.0) 532 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:38.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.591867+0000 mon.vm07 (mon.0) 533 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:38.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.591867+0000 mon.vm07 (mon.0) 533 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:38.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.591892+0000 mon.vm07 (mon.0) 534 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:38.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.591892+0000 mon.vm07 (mon.0) 534 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:38.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.591917+0000 mon.vm07 (mon.0) 535 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:38.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.591917+0000 mon.vm07 (mon.0) 535 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:38.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.591943+0000 mon.vm07 (mon.0) 536 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:38.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.591943+0000 mon.vm07 (mon.0) 536 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:38.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.598127+0000 mon.vm07 (mon.0) 537 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:38.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.598127+0000 mon.vm07 (mon.0) 537 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:38.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.598603+0000 mon.vm07 (mon.0) 538 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:38.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:38 vm07 bash[17031]: audit 2026-03-07T06:55:38.598603+0000 mon.vm07 (mon.0) 538 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:37.748807+0000 mon.vm07 (mon.0) 522 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:37.748807+0000 mon.vm07 (mon.0) 522 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.400860+0000 mon.vm07 (mon.0) 523 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.400860+0000 mon.vm07 (mon.0) 523 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.574473+0000 mon.vm07 (mon.0) 524 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.574473+0000 mon.vm07 (mon.0) 524 : audit [INF] from='osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665]' entity='osd.3' 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.588224+0000 mon.vm07 (mon.0) 525 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.588224+0000 mon.vm07 (mon.0) 525 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.588327+0000 mon.vm07 (mon.0) 526 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.588327+0000 mon.vm07 (mon.0) 526 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.588398+0000 mon.vm07 (mon.0) 527 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.588398+0000 mon.vm07 (mon.0) 527 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: cluster 2026-03-07T06:55:38.591389+0000 mon.vm07 (mon.0) 528 : cluster [DBG] osdmap e18: 8 total, 2 up, 8 in 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: cluster 2026-03-07T06:55:38.591389+0000 mon.vm07 (mon.0) 528 : cluster [DBG] osdmap e18: 8 total, 2 up, 8 in 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.591578+0000 mon.vm07 (mon.0) 529 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.591578+0000 mon.vm07 (mon.0) 529 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.591726+0000 mon.vm07 (mon.0) 530 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.591726+0000 mon.vm07 (mon.0) 530 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.591791+0000 mon.vm07 (mon.0) 531 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.591791+0000 mon.vm07 (mon.0) 531 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.591838+0000 mon.vm07 (mon.0) 532 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.591838+0000 mon.vm07 (mon.0) 532 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.591867+0000 mon.vm07 (mon.0) 533 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.591867+0000 mon.vm07 (mon.0) 533 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.591892+0000 mon.vm07 (mon.0) 534 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.591892+0000 mon.vm07 (mon.0) 534 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.591917+0000 mon.vm07 (mon.0) 535 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.591917+0000 mon.vm07 (mon.0) 535 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.591943+0000 mon.vm07 (mon.0) 536 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.591943+0000 mon.vm07 (mon.0) 536 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.598127+0000 mon.vm07 (mon.0) 537 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.598127+0000 mon.vm07 (mon.0) 537 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.598603+0000 mon.vm07 (mon.0) 538 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:38 vm09 bash[21220]: audit 2026-03-07T06:55:38.598603+0000 mon.vm07 (mon.0) 538 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:39.343 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:55:39.826 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: cluster 2026-03-07T06:55:36.715204+0000 osd.3 (osd.3) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:39.826 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: cluster 2026-03-07T06:55:36.715204+0000 osd.3 (osd.3) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:39.826 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: cluster 2026-03-07T06:55:36.715256+0000 osd.3 (osd.3) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:39.826 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: cluster 2026-03-07T06:55:36.715256+0000 osd.3 (osd.3) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:39.826 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: cluster 2026-03-07T06:55:39.085565+0000 mgr.vm07.yrfcuj (mgr.14201) 90 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-07T07:55:39.827 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: cluster 2026-03-07T06:55:39.085565+0000 mgr.vm07.yrfcuj (mgr.14201) 90 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-07T07:55:39.827 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.262574+0000 mon.vm07 (mon.0) 539 : audit [INF] from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-07T07:55:39.827 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.262574+0000 mon.vm07 (mon.0) 539 : audit [INF] from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-07T07:55:39.827 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.267178+0000 mon.vm09 (mon.1) 14 : audit [INF] from='osd.7 [v2:192.168.123.109:6824/3790896479,v1:192.168.123.109:6825/3790896479]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-07T07:55:39.827 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.267178+0000 mon.vm09 (mon.1) 14 : audit [INF] from='osd.7 [v2:192.168.123.109:6824/3790896479,v1:192.168.123.109:6825/3790896479]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-07T07:55:39.992 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:55:40.098 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":19,"num_osds":8,"num_up_osds":3,"osd_up_since":1772866539,"num_in_osds":8,"osd_in_since":1772866518,"num_remapped_pgs":0} 2026-03-07T07:55:40.098 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.307646+0000 mon.vm07 (mon.0) 540 : audit [INF] from='osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-07T07:55:40.098 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.307646+0000 mon.vm07 (mon.0) 540 : audit [INF] from='osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-07T07:55:40.098 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.575458+0000 mon.vm07 (mon.0) 541 : audit [INF] from='osd.2 ' entity='osd.2' 2026-03-07T07:55:40.098 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.575458+0000 mon.vm07 (mon.0) 541 : audit [INF] from='osd.2 ' entity='osd.2' 2026-03-07T07:55:40.098 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.591021+0000 mon.vm07 (mon.0) 542 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-07T07:55:40.098 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.591021+0000 mon.vm07 (mon.0) 542 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-07T07:55:40.098 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.591138+0000 mon.vm07 (mon.0) 543 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-07T07:55:40.098 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.591138+0000 mon.vm07 (mon.0) 543 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-07T07:55:40.098 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.591169+0000 mon.vm07 (mon.0) 544 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-07T07:55:40.098 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.591169+0000 mon.vm07 (mon.0) 544 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-07T07:55:40.098 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.591202+0000 mon.vm07 (mon.0) 545 : audit [INF] from='osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728]' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-07T07:55:40.098 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.591202+0000 mon.vm07 (mon.0) 545 : audit [INF] from='osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728]' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-07T07:55:40.098 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: cluster 2026-03-07T06:55:39.593349+0000 mon.vm07 (mon.0) 546 : cluster [INF] osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665] boot 2026-03-07T07:55:40.098 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: cluster 2026-03-07T06:55:39.593349+0000 mon.vm07 (mon.0) 546 : cluster [INF] osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665] boot 2026-03-07T07:55:40.098 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: cluster 2026-03-07T06:55:39.593493+0000 mon.vm07 (mon.0) 547 : cluster [DBG] osdmap e19: 8 total, 3 up, 8 in 2026-03-07T07:55:40.098 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: cluster 2026-03-07T06:55:39.593493+0000 mon.vm07 (mon.0) 547 : cluster [DBG] osdmap e19: 8 total, 3 up, 8 in 2026-03-07T07:55:40.098 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.594522+0000 mon.vm07 (mon.0) 548 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:40.098 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.594522+0000 mon.vm07 (mon.0) 548 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:40.098 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.595051+0000 mon.vm07 (mon.0) 549 : audit [INF] from='osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.595051+0000 mon.vm07 (mon.0) 549 : audit [INF] from='osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.595151+0000 mon.vm07 (mon.0) 550 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.595151+0000 mon.vm07 (mon.0) 550 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.595239+0000 mon.vm07 (mon.0) 551 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.595239+0000 mon.vm07 (mon.0) 551 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.595274+0000 mon.vm07 (mon.0) 552 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.595274+0000 mon.vm07 (mon.0) 552 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.595302+0000 mon.vm07 (mon.0) 553 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.595302+0000 mon.vm07 (mon.0) 553 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.595328+0000 mon.vm07 (mon.0) 554 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.595328+0000 mon.vm07 (mon.0) 554 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.600187+0000 mon.vm07 (mon.0) 555 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.600187+0000 mon.vm07 (mon.0) 555 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.601236+0000 mon.vm07 (mon.0) 556 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.601236+0000 mon.vm07 (mon.0) 556 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.602018+0000 mon.vm07 (mon.0) 557 : audit [INF] from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.602018+0000 mon.vm07 (mon.0) 557 : audit [INF] from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.604903+0000 mon.vm07 (mon.0) 558 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.604903+0000 mon.vm07 (mon.0) 558 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.605994+0000 mon.vm09 (mon.1) 15 : audit [INF] from='osd.7 [v2:192.168.123.109:6824/3790896479,v1:192.168.123.109:6825/3790896479]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:40.099 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:39 vm07 bash[17031]: audit 2026-03-07T06:55:39.605994+0000 mon.vm09 (mon.1) 15 : audit [INF] from='osd.7 [v2:192.168.123.109:6824/3790896479,v1:192.168.123.109:6825/3790896479]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:40.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: cluster 2026-03-07T06:55:36.715204+0000 osd.3 (osd.3) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:40.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: cluster 2026-03-07T06:55:36.715204+0000 osd.3 (osd.3) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:40.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: cluster 2026-03-07T06:55:36.715256+0000 osd.3 (osd.3) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:40.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: cluster 2026-03-07T06:55:36.715256+0000 osd.3 (osd.3) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:40.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: cluster 2026-03-07T06:55:39.085565+0000 mgr.vm07.yrfcuj (mgr.14201) 90 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-07T07:55:40.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: cluster 2026-03-07T06:55:39.085565+0000 mgr.vm07.yrfcuj (mgr.14201) 90 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 453 MiB used, 40 GiB / 40 GiB avail 2026-03-07T07:55:40.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.262574+0000 mon.vm07 (mon.0) 539 : audit [INF] from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-07T07:55:40.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.262574+0000 mon.vm07 (mon.0) 539 : audit [INF] from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.267178+0000 mon.vm09 (mon.1) 14 : audit [INF] from='osd.7 [v2:192.168.123.109:6824/3790896479,v1:192.168.123.109:6825/3790896479]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.267178+0000 mon.vm09 (mon.1) 14 : audit [INF] from='osd.7 [v2:192.168.123.109:6824/3790896479,v1:192.168.123.109:6825/3790896479]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.307646+0000 mon.vm07 (mon.0) 540 : audit [INF] from='osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.307646+0000 mon.vm07 (mon.0) 540 : audit [INF] from='osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.575458+0000 mon.vm07 (mon.0) 541 : audit [INF] from='osd.2 ' entity='osd.2' 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.575458+0000 mon.vm07 (mon.0) 541 : audit [INF] from='osd.2 ' entity='osd.2' 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.591021+0000 mon.vm07 (mon.0) 542 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.591021+0000 mon.vm07 (mon.0) 542 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.591138+0000 mon.vm07 (mon.0) 543 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.591138+0000 mon.vm07 (mon.0) 543 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.591169+0000 mon.vm07 (mon.0) 544 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.591169+0000 mon.vm07 (mon.0) 544 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.591202+0000 mon.vm07 (mon.0) 545 : audit [INF] from='osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728]' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.591202+0000 mon.vm07 (mon.0) 545 : audit [INF] from='osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728]' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: cluster 2026-03-07T06:55:39.593349+0000 mon.vm07 (mon.0) 546 : cluster [INF] osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665] boot 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: cluster 2026-03-07T06:55:39.593349+0000 mon.vm07 (mon.0) 546 : cluster [INF] osd.3 [v2:192.168.123.109:6808/1224712665,v1:192.168.123.109:6809/1224712665] boot 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: cluster 2026-03-07T06:55:39.593493+0000 mon.vm07 (mon.0) 547 : cluster [DBG] osdmap e19: 8 total, 3 up, 8 in 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: cluster 2026-03-07T06:55:39.593493+0000 mon.vm07 (mon.0) 547 : cluster [DBG] osdmap e19: 8 total, 3 up, 8 in 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.594522+0000 mon.vm07 (mon.0) 548 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.594522+0000 mon.vm07 (mon.0) 548 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.595051+0000 mon.vm07 (mon.0) 549 : audit [INF] from='osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.595051+0000 mon.vm07 (mon.0) 549 : audit [INF] from='osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.595151+0000 mon.vm07 (mon.0) 550 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.595151+0000 mon.vm07 (mon.0) 550 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.595239+0000 mon.vm07 (mon.0) 551 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.595239+0000 mon.vm07 (mon.0) 551 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.595274+0000 mon.vm07 (mon.0) 552 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.595274+0000 mon.vm07 (mon.0) 552 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.595302+0000 mon.vm07 (mon.0) 553 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.595302+0000 mon.vm07 (mon.0) 553 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.595328+0000 mon.vm07 (mon.0) 554 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.595328+0000 mon.vm07 (mon.0) 554 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.600187+0000 mon.vm07 (mon.0) 555 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.600187+0000 mon.vm07 (mon.0) 555 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.601236+0000 mon.vm07 (mon.0) 556 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.601236+0000 mon.vm07 (mon.0) 556 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.602018+0000 mon.vm07 (mon.0) 557 : audit [INF] from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.602018+0000 mon.vm07 (mon.0) 557 : audit [INF] from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.604903+0000 mon.vm07 (mon.0) 558 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.604903+0000 mon.vm07 (mon.0) 558 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.605994+0000 mon.vm09 (mon.1) 15 : audit [INF] from='osd.7 [v2:192.168.123.109:6824/3790896479,v1:192.168.123.109:6825/3790896479]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:39 vm09 bash[21220]: audit 2026-03-07T06:55:39.605994+0000 mon.vm09 (mon.1) 15 : audit [INF] from='osd.7 [v2:192.168.123.109:6824/3790896479,v1:192.168.123.109:6825/3790896479]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-07T07:55:40.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: cluster 2026-03-07T06:55:37.676615+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:40.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: cluster 2026-03-07T06:55:37.676615+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:40.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: cluster 2026-03-07T06:55:37.676668+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:40.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: cluster 2026-03-07T06:55:37.676668+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:40.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: cluster 2026-03-07T06:55:38.752660+0000 osd.4 (osd.4) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:40.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: cluster 2026-03-07T06:55:38.752660+0000 osd.4 (osd.4) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:40.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: cluster 2026-03-07T06:55:38.752712+0000 osd.4 (osd.4) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:40.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: cluster 2026-03-07T06:55:38.752712+0000 osd.4 (osd.4) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:40.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:39.986412+0000 mon.vm07 (mon.0) 559 : audit [DBG] from='client.? 192.168.123.107:0/2608184123' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:39.986412+0000 mon.vm07 (mon.0) 559 : audit [DBG] from='client.? 192.168.123.107:0/2608184123' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.124949+0000 mon.vm07 (mon.0) 560 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.124949+0000 mon.vm07 (mon.0) 560 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.598323+0000 mon.vm07 (mon.0) 561 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.598323+0000 mon.vm07 (mon.0) 561 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.599430+0000 mon.vm07 (mon.0) 562 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.599430+0000 mon.vm07 (mon.0) 562 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.599889+0000 mon.vm07 (mon.0) 563 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.599889+0000 mon.vm07 (mon.0) 563 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.609334+0000 mon.vm07 (mon.0) 564 : audit [INF] from='osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728]' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.609334+0000 mon.vm07 (mon.0) 564 : audit [INF] from='osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728]' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.609467+0000 mon.vm07 (mon.0) 565 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.609467+0000 mon.vm07 (mon.0) 565 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: cluster 2026-03-07T06:55:40.614199+0000 mon.vm07 (mon.0) 566 : cluster [INF] osd.2 [v2:192.168.123.107:6810/158678589,v1:192.168.123.107:6811/158678589] boot 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: cluster 2026-03-07T06:55:40.614199+0000 mon.vm07 (mon.0) 566 : cluster [INF] osd.2 [v2:192.168.123.107:6810/158678589,v1:192.168.123.107:6811/158678589] boot 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: cluster 2026-03-07T06:55:40.614321+0000 mon.vm07 (mon.0) 567 : cluster [DBG] osdmap e20: 8 total, 4 up, 8 in 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: cluster 2026-03-07T06:55:40.614321+0000 mon.vm07 (mon.0) 567 : cluster [DBG] osdmap e20: 8 total, 4 up, 8 in 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.615364+0000 mon.vm07 (mon.0) 568 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.615364+0000 mon.vm07 (mon.0) 568 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.615803+0000 mon.vm07 (mon.0) 569 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.615803+0000 mon.vm07 (mon.0) 569 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.616055+0000 mon.vm07 (mon.0) 570 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.616055+0000 mon.vm07 (mon.0) 570 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.616211+0000 mon.vm07 (mon.0) 571 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.616211+0000 mon.vm07 (mon.0) 571 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.616369+0000 mon.vm07 (mon.0) 572 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.616369+0000 mon.vm07 (mon.0) 572 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.618917+0000 mon.vm07 (mon.0) 573 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.618917+0000 mon.vm07 (mon.0) 573 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.620533+0000 mon.vm07 (mon.0) 574 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:40.909 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:40 vm07 bash[17031]: audit 2026-03-07T06:55:40.620533+0000 mon.vm07 (mon.0) 574 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:41.099 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd stat -f json 2026-03-07T07:55:41.128 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: cluster 2026-03-07T06:55:37.676615+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: cluster 2026-03-07T06:55:37.676615+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: cluster 2026-03-07T06:55:37.676668+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: cluster 2026-03-07T06:55:37.676668+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: cluster 2026-03-07T06:55:38.752660+0000 osd.4 (osd.4) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: cluster 2026-03-07T06:55:38.752660+0000 osd.4 (osd.4) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: cluster 2026-03-07T06:55:38.752712+0000 osd.4 (osd.4) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: cluster 2026-03-07T06:55:38.752712+0000 osd.4 (osd.4) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:39.986412+0000 mon.vm07 (mon.0) 559 : audit [DBG] from='client.? 192.168.123.107:0/2608184123' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:39.986412+0000 mon.vm07 (mon.0) 559 : audit [DBG] from='client.? 192.168.123.107:0/2608184123' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.124949+0000 mon.vm07 (mon.0) 560 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.124949+0000 mon.vm07 (mon.0) 560 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.598323+0000 mon.vm07 (mon.0) 561 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.598323+0000 mon.vm07 (mon.0) 561 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.599430+0000 mon.vm07 (mon.0) 562 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.599430+0000 mon.vm07 (mon.0) 562 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.599889+0000 mon.vm07 (mon.0) 563 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.599889+0000 mon.vm07 (mon.0) 563 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.609334+0000 mon.vm07 (mon.0) 564 : audit [INF] from='osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728]' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.609334+0000 mon.vm07 (mon.0) 564 : audit [INF] from='osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728]' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.609467+0000 mon.vm07 (mon.0) 565 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.609467+0000 mon.vm07 (mon.0) 565 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: cluster 2026-03-07T06:55:40.614199+0000 mon.vm07 (mon.0) 566 : cluster [INF] osd.2 [v2:192.168.123.107:6810/158678589,v1:192.168.123.107:6811/158678589] boot 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: cluster 2026-03-07T06:55:40.614199+0000 mon.vm07 (mon.0) 566 : cluster [INF] osd.2 [v2:192.168.123.107:6810/158678589,v1:192.168.123.107:6811/158678589] boot 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: cluster 2026-03-07T06:55:40.614321+0000 mon.vm07 (mon.0) 567 : cluster [DBG] osdmap e20: 8 total, 4 up, 8 in 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: cluster 2026-03-07T06:55:40.614321+0000 mon.vm07 (mon.0) 567 : cluster [DBG] osdmap e20: 8 total, 4 up, 8 in 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.615364+0000 mon.vm07 (mon.0) 568 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.615364+0000 mon.vm07 (mon.0) 568 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.615803+0000 mon.vm07 (mon.0) 569 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.615803+0000 mon.vm07 (mon.0) 569 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.616055+0000 mon.vm07 (mon.0) 570 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.616055+0000 mon.vm07 (mon.0) 570 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.616211+0000 mon.vm07 (mon.0) 571 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.616211+0000 mon.vm07 (mon.0) 571 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.616369+0000 mon.vm07 (mon.0) 572 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.616369+0000 mon.vm07 (mon.0) 572 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.618917+0000 mon.vm07 (mon.0) 573 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.618917+0000 mon.vm07 (mon.0) 573 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.620533+0000 mon.vm07 (mon.0) 574 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:41.129 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:40 vm09 bash[21220]: audit 2026-03-07T06:55:40.620533+0000 mon.vm07 (mon.0) 574 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:42.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: cluster 2026-03-07T06:55:39.447105+0000 osd.5 (osd.5) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:42.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: cluster 2026-03-07T06:55:39.447105+0000 osd.5 (osd.5) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:42.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: cluster 2026-03-07T06:55:39.447155+0000 osd.5 (osd.5) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: cluster 2026-03-07T06:55:39.447155+0000 osd.5 (osd.5) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: cluster 2026-03-07T06:55:41.086992+0000 mgr.vm07.yrfcuj (mgr.14201) 91 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 479 MiB used, 60 GiB / 60 GiB avail 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: cluster 2026-03-07T06:55:41.086992+0000 mgr.vm07.yrfcuj (mgr.14201) 91 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 479 MiB used, 60 GiB / 60 GiB avail 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.130059+0000 mon.vm07 (mon.0) 575 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.130059+0000 mon.vm07 (mon.0) 575 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.203063+0000 mon.vm07 (mon.0) 576 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.203063+0000 mon.vm07 (mon.0) 576 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.225727+0000 mon.vm07 (mon.0) 577 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.225727+0000 mon.vm07 (mon.0) 577 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.599183+0000 mon.vm07 (mon.0) 578 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.599183+0000 mon.vm07 (mon.0) 578 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.600048+0000 mon.vm07 (mon.0) 579 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.600048+0000 mon.vm07 (mon.0) 579 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.612618+0000 mon.vm07 (mon.0) 580 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.612618+0000 mon.vm07 (mon.0) 580 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: cluster 2026-03-07T06:55:41.614615+0000 mon.vm07 (mon.0) 581 : cluster [INF] osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728] boot 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: cluster 2026-03-07T06:55:41.614615+0000 mon.vm07 (mon.0) 581 : cluster [INF] osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728] boot 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: cluster 2026-03-07T06:55:41.614628+0000 mon.vm07 (mon.0) 582 : cluster [INF] osd.7 [v2:192.168.123.109:6824/3790896479,v1:192.168.123.109:6825/3790896479] boot 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: cluster 2026-03-07T06:55:41.614628+0000 mon.vm07 (mon.0) 582 : cluster [INF] osd.7 [v2:192.168.123.109:6824/3790896479,v1:192.168.123.109:6825/3790896479] boot 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: cluster 2026-03-07T06:55:41.614640+0000 mon.vm07 (mon.0) 583 : cluster [INF] osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676] boot 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: cluster 2026-03-07T06:55:41.614640+0000 mon.vm07 (mon.0) 583 : cluster [INF] osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676] boot 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: cluster 2026-03-07T06:55:41.614654+0000 mon.vm07 (mon.0) 584 : cluster [INF] osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023] boot 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: cluster 2026-03-07T06:55:41.614654+0000 mon.vm07 (mon.0) 584 : cluster [INF] osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023] boot 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: cluster 2026-03-07T06:55:41.614667+0000 mon.vm07 (mon.0) 585 : cluster [DBG] osdmap e21: 8 total, 8 up, 8 in 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: cluster 2026-03-07T06:55:41.614667+0000 mon.vm07 (mon.0) 585 : cluster [DBG] osdmap e21: 8 total, 8 up, 8 in 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.615004+0000 mon.vm07 (mon.0) 586 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.615004+0000 mon.vm07 (mon.0) 586 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.615087+0000 mon.vm07 (mon.0) 587 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.615087+0000 mon.vm07 (mon.0) 587 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.615138+0000 mon.vm07 (mon.0) 588 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.615138+0000 mon.vm07 (mon.0) 588 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.615186+0000 mon.vm07 (mon.0) 589 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.615186+0000 mon.vm07 (mon.0) 589 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.625160+0000 mon.vm07 (mon.0) 590 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-07T07:55:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:41 vm09 bash[21220]: audit 2026-03-07T06:55:41.625160+0000 mon.vm07 (mon.0) 590 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: cluster 2026-03-07T06:55:39.447105+0000 osd.5 (osd.5) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: cluster 2026-03-07T06:55:39.447105+0000 osd.5 (osd.5) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: cluster 2026-03-07T06:55:39.447155+0000 osd.5 (osd.5) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: cluster 2026-03-07T06:55:39.447155+0000 osd.5 (osd.5) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: cluster 2026-03-07T06:55:41.086992+0000 mgr.vm07.yrfcuj (mgr.14201) 91 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 479 MiB used, 60 GiB / 60 GiB avail 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: cluster 2026-03-07T06:55:41.086992+0000 mgr.vm07.yrfcuj (mgr.14201) 91 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 479 MiB used, 60 GiB / 60 GiB avail 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.130059+0000 mon.vm07 (mon.0) 575 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.130059+0000 mon.vm07 (mon.0) 575 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.203063+0000 mon.vm07 (mon.0) 576 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.203063+0000 mon.vm07 (mon.0) 576 : audit [INF] from='osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676]' entity='osd.5' 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.225727+0000 mon.vm07 (mon.0) 577 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.225727+0000 mon.vm07 (mon.0) 577 : audit [INF] from='osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023]' entity='osd.4' 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.599183+0000 mon.vm07 (mon.0) 578 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.599183+0000 mon.vm07 (mon.0) 578 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.600048+0000 mon.vm07 (mon.0) 579 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.600048+0000 mon.vm07 (mon.0) 579 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.612618+0000 mon.vm07 (mon.0) 580 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.612618+0000 mon.vm07 (mon.0) 580 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: cluster 2026-03-07T06:55:41.614615+0000 mon.vm07 (mon.0) 581 : cluster [INF] osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728] boot 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: cluster 2026-03-07T06:55:41.614615+0000 mon.vm07 (mon.0) 581 : cluster [INF] osd.6 [v2:192.168.123.107:6826/2603438728,v1:192.168.123.107:6827/2603438728] boot 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: cluster 2026-03-07T06:55:41.614628+0000 mon.vm07 (mon.0) 582 : cluster [INF] osd.7 [v2:192.168.123.109:6824/3790896479,v1:192.168.123.109:6825/3790896479] boot 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: cluster 2026-03-07T06:55:41.614628+0000 mon.vm07 (mon.0) 582 : cluster [INF] osd.7 [v2:192.168.123.109:6824/3790896479,v1:192.168.123.109:6825/3790896479] boot 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: cluster 2026-03-07T06:55:41.614640+0000 mon.vm07 (mon.0) 583 : cluster [INF] osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676] boot 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: cluster 2026-03-07T06:55:41.614640+0000 mon.vm07 (mon.0) 583 : cluster [INF] osd.5 [v2:192.168.123.107:6818/212397676,v1:192.168.123.107:6819/212397676] boot 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: cluster 2026-03-07T06:55:41.614654+0000 mon.vm07 (mon.0) 584 : cluster [INF] osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023] boot 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: cluster 2026-03-07T06:55:41.614654+0000 mon.vm07 (mon.0) 584 : cluster [INF] osd.4 [v2:192.168.123.109:6816/1602816023,v1:192.168.123.109:6817/1602816023] boot 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: cluster 2026-03-07T06:55:41.614667+0000 mon.vm07 (mon.0) 585 : cluster [DBG] osdmap e21: 8 total, 8 up, 8 in 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: cluster 2026-03-07T06:55:41.614667+0000 mon.vm07 (mon.0) 585 : cluster [DBG] osdmap e21: 8 total, 8 up, 8 in 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.615004+0000 mon.vm07 (mon.0) 586 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:42.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.615004+0000 mon.vm07 (mon.0) 586 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-07T07:55:42.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.615087+0000 mon.vm07 (mon.0) 587 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:42.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.615087+0000 mon.vm07 (mon.0) 587 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-07T07:55:42.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.615138+0000 mon.vm07 (mon.0) 588 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:42.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.615138+0000 mon.vm07 (mon.0) 588 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-07T07:55:42.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.615186+0000 mon.vm07 (mon.0) 589 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:42.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.615186+0000 mon.vm07 (mon.0) 589 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-07T07:55:42.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.625160+0000 mon.vm07 (mon.0) 590 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-07T07:55:42.159 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:41 vm07 bash[17031]: audit 2026-03-07T06:55:41.625160+0000 mon.vm07 (mon.0) 590 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-07T07:55:43.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:42 vm09 bash[21220]: cluster 2026-03-07T06:55:40.260961+0000 osd.7 (osd.7) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:43.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:42 vm09 bash[21220]: cluster 2026-03-07T06:55:40.260961+0000 osd.7 (osd.7) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:43.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:42 vm09 bash[21220]: cluster 2026-03-07T06:55:40.261047+0000 osd.7 (osd.7) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:43.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:42 vm09 bash[21220]: cluster 2026-03-07T06:55:40.261047+0000 osd.7 (osd.7) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:43.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:42 vm09 bash[21220]: cluster 2026-03-07T06:55:40.346093+0000 osd.6 (osd.6) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:43.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:42 vm09 bash[21220]: cluster 2026-03-07T06:55:40.346093+0000 osd.6 (osd.6) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:43.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:42 vm09 bash[21220]: cluster 2026-03-07T06:55:40.346149+0000 osd.6 (osd.6) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:43.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:42 vm09 bash[21220]: cluster 2026-03-07T06:55:40.346149+0000 osd.6 (osd.6) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:43.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:42 vm09 bash[21220]: audit 2026-03-07T06:55:42.616260+0000 mon.vm07 (mon.0) 591 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-07T07:55:43.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:42 vm09 bash[21220]: audit 2026-03-07T06:55:42.616260+0000 mon.vm07 (mon.0) 591 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-07T07:55:43.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:42 vm09 bash[21220]: cluster 2026-03-07T06:55:42.619104+0000 mon.vm07 (mon.0) 592 : cluster [DBG] osdmap e22: 8 total, 8 up, 8 in 2026-03-07T07:55:43.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:42 vm09 bash[21220]: cluster 2026-03-07T06:55:42.619104+0000 mon.vm07 (mon.0) 592 : cluster [DBG] osdmap e22: 8 total, 8 up, 8 in 2026-03-07T07:55:43.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:42 vm07 bash[17031]: cluster 2026-03-07T06:55:40.260961+0000 osd.7 (osd.7) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:43.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:42 vm07 bash[17031]: cluster 2026-03-07T06:55:40.260961+0000 osd.7 (osd.7) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:43.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:42 vm07 bash[17031]: cluster 2026-03-07T06:55:40.261047+0000 osd.7 (osd.7) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:43.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:42 vm07 bash[17031]: cluster 2026-03-07T06:55:40.261047+0000 osd.7 (osd.7) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:43.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:42 vm07 bash[17031]: cluster 2026-03-07T06:55:40.346093+0000 osd.6 (osd.6) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:43.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:42 vm07 bash[17031]: cluster 2026-03-07T06:55:40.346093+0000 osd.6 (osd.6) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-07T07:55:43.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:42 vm07 bash[17031]: cluster 2026-03-07T06:55:40.346149+0000 osd.6 (osd.6) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:43.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:42 vm07 bash[17031]: cluster 2026-03-07T06:55:40.346149+0000 osd.6 (osd.6) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-07T07:55:43.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:42 vm07 bash[17031]: audit 2026-03-07T06:55:42.616260+0000 mon.vm07 (mon.0) 591 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-07T07:55:43.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:42 vm07 bash[17031]: audit 2026-03-07T06:55:42.616260+0000 mon.vm07 (mon.0) 591 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-07T07:55:43.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:42 vm07 bash[17031]: cluster 2026-03-07T06:55:42.619104+0000 mon.vm07 (mon.0) 592 : cluster [DBG] osdmap e22: 8 total, 8 up, 8 in 2026-03-07T07:55:43.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:42 vm07 bash[17031]: cluster 2026-03-07T06:55:42.619104+0000 mon.vm07 (mon.0) 592 : cluster [DBG] osdmap e22: 8 total, 8 up, 8 in 2026-03-07T07:55:44.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: cluster 2026-03-07T06:55:43.087301+0000 mgr.vm07.yrfcuj (mgr.14201) 92 : cluster [DBG] pgmap v49: 1 pgs: 1 creating+peering; 0 B data, 611 MiB used, 159 GiB / 160 GiB avail 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: cluster 2026-03-07T06:55:43.087301+0000 mgr.vm07.yrfcuj (mgr.14201) 92 : cluster [DBG] pgmap v49: 1 pgs: 1 creating+peering; 0 B data, 611 MiB used, 159 GiB / 160 GiB avail 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.206546+0000 mon.vm07 (mon.0) 593 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.206546+0000 mon.vm07 (mon.0) 593 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.211737+0000 mon.vm07 (mon.0) 594 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.211737+0000 mon.vm07 (mon.0) 594 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.236824+0000 mon.vm07 (mon.0) 595 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.236824+0000 mon.vm07 (mon.0) 595 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.241872+0000 mon.vm07 (mon.0) 596 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.241872+0000 mon.vm07 (mon.0) 596 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.287798+0000 mon.vm07 (mon.0) 597 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.287798+0000 mon.vm07 (mon.0) 597 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.480741+0000 mon.vm07 (mon.0) 598 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.480741+0000 mon.vm07 (mon.0) 598 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.498406+0000 mon.vm07 (mon.0) 599 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.498406+0000 mon.vm07 (mon.0) 599 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.498595+0000 mon.vm07 (mon.0) 600 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.498595+0000 mon.vm07 (mon.0) 600 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.498674+0000 mon.vm07 (mon.0) 601 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.498674+0000 mon.vm07 (mon.0) 601 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.500889+0000 mon.vm07 (mon.0) 602 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.500889+0000 mon.vm07 (mon.0) 602 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.500975+0000 mon.vm07 (mon.0) 603 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.500975+0000 mon.vm07 (mon.0) 603 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.505700+0000 mon.vm09 (mon.1) 16 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.505700+0000 mon.vm09 (mon.1) 16 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.522640+0000 mon.vm09 (mon.1) 17 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-07T07:55:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:44 vm09 bash[21220]: audit 2026-03-07T06:55:43.522640+0000 mon.vm09 (mon.1) 17 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: cluster 2026-03-07T06:55:43.087301+0000 mgr.vm07.yrfcuj (mgr.14201) 92 : cluster [DBG] pgmap v49: 1 pgs: 1 creating+peering; 0 B data, 611 MiB used, 159 GiB / 160 GiB avail 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: cluster 2026-03-07T06:55:43.087301+0000 mgr.vm07.yrfcuj (mgr.14201) 92 : cluster [DBG] pgmap v49: 1 pgs: 1 creating+peering; 0 B data, 611 MiB used, 159 GiB / 160 GiB avail 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.206546+0000 mon.vm07 (mon.0) 593 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.206546+0000 mon.vm07 (mon.0) 593 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.211737+0000 mon.vm07 (mon.0) 594 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.211737+0000 mon.vm07 (mon.0) 594 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.236824+0000 mon.vm07 (mon.0) 595 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.236824+0000 mon.vm07 (mon.0) 595 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.241872+0000 mon.vm07 (mon.0) 596 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.241872+0000 mon.vm07 (mon.0) 596 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.287798+0000 mon.vm07 (mon.0) 597 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.287798+0000 mon.vm07 (mon.0) 597 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.480741+0000 mon.vm07 (mon.0) 598 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.480741+0000 mon.vm07 (mon.0) 598 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.498406+0000 mon.vm07 (mon.0) 599 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.498406+0000 mon.vm07 (mon.0) 599 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.498595+0000 mon.vm07 (mon.0) 600 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.498595+0000 mon.vm07 (mon.0) 600 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.498674+0000 mon.vm07 (mon.0) 601 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.498674+0000 mon.vm07 (mon.0) 601 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.500889+0000 mon.vm07 (mon.0) 602 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.500889+0000 mon.vm07 (mon.0) 602 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm07"}]: dispatch 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.500975+0000 mon.vm07 (mon.0) 603 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.500975+0000 mon.vm07 (mon.0) 603 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.505700+0000 mon.vm09 (mon.1) 16 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.505700+0000 mon.vm09 (mon.1) 16 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.522640+0000 mon.vm09 (mon.1) 17 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-07T07:55:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:44 vm07 bash[17031]: audit 2026-03-07T06:55:43.522640+0000 mon.vm09 (mon.1) 17 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-07T07:55:45.399 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:55:45.482 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:45 vm07 bash[17031]: cluster 2026-03-07T06:55:44.245284+0000 mon.vm07 (mon.0) 604 : cluster [DBG] osdmap e23: 8 total, 8 up, 8 in 2026-03-07T07:55:45.482 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:45 vm07 bash[17031]: cluster 2026-03-07T06:55:44.245284+0000 mon.vm07 (mon.0) 604 : cluster [DBG] osdmap e23: 8 total, 8 up, 8 in 2026-03-07T07:55:45.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:45 vm09 bash[21220]: cluster 2026-03-07T06:55:44.245284+0000 mon.vm07 (mon.0) 604 : cluster [DBG] osdmap e23: 8 total, 8 up, 8 in 2026-03-07T07:55:45.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:45 vm09 bash[21220]: cluster 2026-03-07T06:55:44.245284+0000 mon.vm07 (mon.0) 604 : cluster [DBG] osdmap e23: 8 total, 8 up, 8 in 2026-03-07T07:55:45.737 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:55:45.793 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":23,"num_osds":8,"num_up_osds":8,"osd_up_since":1772866541,"num_in_osds":8,"osd_in_since":1772866518,"num_remapped_pgs":0} 2026-03-07T07:55:45.793 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd dump --format=json 2026-03-07T07:55:46.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:46 vm09 bash[21220]: cluster 2026-03-07T06:55:45.087569+0000 mgr.vm07.yrfcuj (mgr.14201) 93 : cluster [DBG] pgmap v51: 1 pgs: 1 creating+peering; 0 B data, 612 MiB used, 159 GiB / 160 GiB avail 2026-03-07T07:55:46.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:46 vm09 bash[21220]: cluster 2026-03-07T06:55:45.087569+0000 mgr.vm07.yrfcuj (mgr.14201) 93 : cluster [DBG] pgmap v51: 1 pgs: 1 creating+peering; 0 B data, 612 MiB used, 159 GiB / 160 GiB avail 2026-03-07T07:55:46.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:46 vm09 bash[21220]: cluster 2026-03-07T06:55:45.260273+0000 mon.vm07 (mon.0) 605 : cluster [DBG] mgrmap e19: vm07.yrfcuj(active, since 80s), standbys: vm09.eqznpw 2026-03-07T07:55:46.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:46 vm09 bash[21220]: cluster 2026-03-07T06:55:45.260273+0000 mon.vm07 (mon.0) 605 : cluster [DBG] mgrmap e19: vm07.yrfcuj(active, since 80s), standbys: vm09.eqznpw 2026-03-07T07:55:46.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:46 vm09 bash[21220]: audit 2026-03-07T06:55:45.731549+0000 mon.vm07 (mon.0) 606 : audit [DBG] from='client.? 192.168.123.107:0/2855795536' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:46.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:46 vm09 bash[21220]: audit 2026-03-07T06:55:45.731549+0000 mon.vm07 (mon.0) 606 : audit [DBG] from='client.? 192.168.123.107:0/2855795536' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:46.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:46 vm07 bash[17031]: cluster 2026-03-07T06:55:45.087569+0000 mgr.vm07.yrfcuj (mgr.14201) 93 : cluster [DBG] pgmap v51: 1 pgs: 1 creating+peering; 0 B data, 612 MiB used, 159 GiB / 160 GiB avail 2026-03-07T07:55:46.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:46 vm07 bash[17031]: cluster 2026-03-07T06:55:45.087569+0000 mgr.vm07.yrfcuj (mgr.14201) 93 : cluster [DBG] pgmap v51: 1 pgs: 1 creating+peering; 0 B data, 612 MiB used, 159 GiB / 160 GiB avail 2026-03-07T07:55:46.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:46 vm07 bash[17031]: cluster 2026-03-07T06:55:45.260273+0000 mon.vm07 (mon.0) 605 : cluster [DBG] mgrmap e19: vm07.yrfcuj(active, since 80s), standbys: vm09.eqznpw 2026-03-07T07:55:46.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:46 vm07 bash[17031]: cluster 2026-03-07T06:55:45.260273+0000 mon.vm07 (mon.0) 605 : cluster [DBG] mgrmap e19: vm07.yrfcuj(active, since 80s), standbys: vm09.eqznpw 2026-03-07T07:55:46.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:46 vm07 bash[17031]: audit 2026-03-07T06:55:45.731549+0000 mon.vm07 (mon.0) 606 : audit [DBG] from='client.? 192.168.123.107:0/2855795536' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:46.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:46 vm07 bash[17031]: audit 2026-03-07T06:55:45.731549+0000 mon.vm07 (mon.0) 606 : audit [DBG] from='client.? 192.168.123.107:0/2855795536' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-07T07:55:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:48 vm09 bash[21220]: cluster 2026-03-07T06:55:47.091074+0000 mgr.vm07.yrfcuj (mgr.14201) 94 : cluster [DBG] pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 612 MiB used, 159 GiB / 160 GiB avail 2026-03-07T07:55:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:48 vm09 bash[21220]: cluster 2026-03-07T06:55:47.091074+0000 mgr.vm07.yrfcuj (mgr.14201) 94 : cluster [DBG] pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 612 MiB used, 159 GiB / 160 GiB avail 2026-03-07T07:55:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:48 vm09 bash[21220]: audit 2026-03-07T06:55:48.062058+0000 mon.vm07 (mon.0) 607 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:48 vm09 bash[21220]: audit 2026-03-07T06:55:48.062058+0000 mon.vm07 (mon.0) 607 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:48 vm09 bash[21220]: audit 2026-03-07T06:55:48.107970+0000 mon.vm07 (mon.0) 608 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:48 vm09 bash[21220]: audit 2026-03-07T06:55:48.107970+0000 mon.vm07 (mon.0) 608 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:48 vm07 bash[17031]: cluster 2026-03-07T06:55:47.091074+0000 mgr.vm07.yrfcuj (mgr.14201) 94 : cluster [DBG] pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 612 MiB used, 159 GiB / 160 GiB avail 2026-03-07T07:55:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:48 vm07 bash[17031]: cluster 2026-03-07T06:55:47.091074+0000 mgr.vm07.yrfcuj (mgr.14201) 94 : cluster [DBG] pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 612 MiB used, 159 GiB / 160 GiB avail 2026-03-07T07:55:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:48 vm07 bash[17031]: audit 2026-03-07T06:55:48.062058+0000 mon.vm07 (mon.0) 607 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:48 vm07 bash[17031]: audit 2026-03-07T06:55:48.062058+0000 mon.vm07 (mon.0) 607 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:48 vm07 bash[17031]: audit 2026-03-07T06:55:48.107970+0000 mon.vm07 (mon.0) 608 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:48 vm07 bash[17031]: audit 2026-03-07T06:55:48.107970+0000 mon.vm07 (mon.0) 608 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:49.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:49 vm09 bash[21220]: audit 2026-03-07T06:55:48.504408+0000 mon.vm07 (mon.0) 609 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:49.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:49 vm09 bash[21220]: audit 2026-03-07T06:55:48.504408+0000 mon.vm07 (mon.0) 609 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:49.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:49 vm09 bash[21220]: audit 2026-03-07T06:55:48.509008+0000 mon.vm07 (mon.0) 610 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:49.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:49 vm09 bash[21220]: audit 2026-03-07T06:55:48.509008+0000 mon.vm07 (mon.0) 610 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:49.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:49 vm09 bash[21220]: cluster 2026-03-07T06:55:49.091404+0000 mgr.vm07.yrfcuj (mgr.14201) 95 : cluster [DBG] pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:55:49.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:49 vm09 bash[21220]: cluster 2026-03-07T06:55:49.091404+0000 mgr.vm07.yrfcuj (mgr.14201) 95 : cluster [DBG] pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:55:49.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:49 vm07 bash[17031]: audit 2026-03-07T06:55:48.504408+0000 mon.vm07 (mon.0) 609 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:49.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:49 vm07 bash[17031]: audit 2026-03-07T06:55:48.504408+0000 mon.vm07 (mon.0) 609 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:49.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:49 vm07 bash[17031]: audit 2026-03-07T06:55:48.509008+0000 mon.vm07 (mon.0) 610 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:49.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:49 vm07 bash[17031]: audit 2026-03-07T06:55:48.509008+0000 mon.vm07 (mon.0) 610 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:49.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:49 vm07 bash[17031]: cluster 2026-03-07T06:55:49.091404+0000 mgr.vm07.yrfcuj (mgr.14201) 95 : cluster [DBG] pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:55:49.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:49 vm07 bash[17031]: cluster 2026-03-07T06:55:49.091404+0000 mgr.vm07.yrfcuj (mgr.14201) 95 : cluster [DBG] pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:55:50.445 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:55:50.787 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:55:50.787 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":23,"fsid":"312fdbc4-19f2-11f1-81d0-4bbd10a1e012","created":"2026-03-07T06:52:52.038690+0000","modified":"2026-03-07T06:55:44.240571+0000","last_up_change":"2026-03-07T06:55:41.609166+0000","last_in_change":"2026-03-07T06:55:18.145593+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":9,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-07T06:55:41.135780+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"23","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"e6429ca8-de99-4811-a21e-2292aedf25ef","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":635544560},{"type":"v1","addr":"192.168.123.109:6801","nonce":635544560}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":635544560},{"type":"v1","addr":"192.168.123.109:6803","nonce":635544560}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":635544560},{"type":"v1","addr":"192.168.123.109:6807","nonce":635544560}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":635544560},{"type":"v1","addr":"192.168.123.109:6805","nonce":635544560}]},"public_addr":"192.168.123.109:6801/635544560","cluster_addr":"192.168.123.109:6803/635544560","heartbeat_back_addr":"192.168.123.109:6807/635544560","heartbeat_front_addr":"192.168.123.109:6805/635544560","state":["exists","up"]},{"osd":1,"uuid":"56cba12f-96c0-445c-be10-c84fa665caa4","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6802","nonce":1359088522},{"type":"v1","addr":"192.168.123.107:6803","nonce":1359088522}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6804","nonce":1359088522},{"type":"v1","addr":"192.168.123.107:6805","nonce":1359088522}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6808","nonce":1359088522},{"type":"v1","addr":"192.168.123.107:6809","nonce":1359088522}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6806","nonce":1359088522},{"type":"v1","addr":"192.168.123.107:6807","nonce":1359088522}]},"public_addr":"192.168.123.107:6803/1359088522","cluster_addr":"192.168.123.107:6805/1359088522","heartbeat_back_addr":"192.168.123.107:6809/1359088522","heartbeat_front_addr":"192.168.123.107:6807/1359088522","state":["exists","up"]},{"osd":2,"uuid":"2a313541-d79f-4e9b-8bb5-038ec51951db","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6810","nonce":158678589},{"type":"v1","addr":"192.168.123.107:6811","nonce":158678589}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6812","nonce":158678589},{"type":"v1","addr":"192.168.123.107:6813","nonce":158678589}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6816","nonce":158678589},{"type":"v1","addr":"192.168.123.107:6817","nonce":158678589}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6814","nonce":158678589},{"type":"v1","addr":"192.168.123.107:6815","nonce":158678589}]},"public_addr":"192.168.123.107:6811/158678589","cluster_addr":"192.168.123.107:6813/158678589","heartbeat_back_addr":"192.168.123.107:6817/158678589","heartbeat_front_addr":"192.168.123.107:6815/158678589","state":["exists","up"]},{"osd":3,"uuid":"a583cddd-35d0-41d0-83c4-877b3c089e8b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":19,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6808","nonce":1224712665},{"type":"v1","addr":"192.168.123.109:6809","nonce":1224712665}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6810","nonce":1224712665},{"type":"v1","addr":"192.168.123.109:6811","nonce":1224712665}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6814","nonce":1224712665},{"type":"v1","addr":"192.168.123.109:6815","nonce":1224712665}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6812","nonce":1224712665},{"type":"v1","addr":"192.168.123.109:6813","nonce":1224712665}]},"public_addr":"192.168.123.109:6809/1224712665","cluster_addr":"192.168.123.109:6811/1224712665","heartbeat_back_addr":"192.168.123.109:6815/1224712665","heartbeat_front_addr":"192.168.123.109:6813/1224712665","state":["exists","up"]},{"osd":4,"uuid":"2f651ee5-5a7b-4d0c-82f4-907f0a8beddc","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6816","nonce":1602816023},{"type":"v1","addr":"192.168.123.109:6817","nonce":1602816023}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6818","nonce":1602816023},{"type":"v1","addr":"192.168.123.109:6819","nonce":1602816023}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6822","nonce":1602816023},{"type":"v1","addr":"192.168.123.109:6823","nonce":1602816023}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6820","nonce":1602816023},{"type":"v1","addr":"192.168.123.109:6821","nonce":1602816023}]},"public_addr":"192.168.123.109:6817/1602816023","cluster_addr":"192.168.123.109:6819/1602816023","heartbeat_back_addr":"192.168.123.109:6823/1602816023","heartbeat_front_addr":"192.168.123.109:6821/1602816023","state":["exists","up"]},{"osd":5,"uuid":"5874e093-78e8-4710-8cf9-d27b73d59b9b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6818","nonce":212397676},{"type":"v1","addr":"192.168.123.107:6819","nonce":212397676}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6820","nonce":212397676},{"type":"v1","addr":"192.168.123.107:6821","nonce":212397676}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6824","nonce":212397676},{"type":"v1","addr":"192.168.123.107:6825","nonce":212397676}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6822","nonce":212397676},{"type":"v1","addr":"192.168.123.107:6823","nonce":212397676}]},"public_addr":"192.168.123.107:6819/212397676","cluster_addr":"192.168.123.107:6821/212397676","heartbeat_back_addr":"192.168.123.107:6825/212397676","heartbeat_front_addr":"192.168.123.107:6823/212397676","state":["exists","up"]},{"osd":6,"uuid":"85f57ef2-7375-4150-b632-6d08ce1f021e","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6826","nonce":2603438728},{"type":"v1","addr":"192.168.123.107:6827","nonce":2603438728}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6828","nonce":2603438728},{"type":"v1","addr":"192.168.123.107:6829","nonce":2603438728}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6832","nonce":2603438728},{"type":"v1","addr":"192.168.123.107:6833","nonce":2603438728}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6830","nonce":2603438728},{"type":"v1","addr":"192.168.123.107:6831","nonce":2603438728}]},"public_addr":"192.168.123.107:6827/2603438728","cluster_addr":"192.168.123.107:6829/2603438728","heartbeat_back_addr":"192.168.123.107:6833/2603438728","heartbeat_front_addr":"192.168.123.107:6831/2603438728","state":["exists","up"]},{"osd":7,"uuid":"6ea2bf16-f584-49a8-ba45-a4b9ba5043ce","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":21,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6824","nonce":3790896479},{"type":"v1","addr":"192.168.123.109:6825","nonce":3790896479}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6826","nonce":3790896479},{"type":"v1","addr":"192.168.123.109:6827","nonce":3790896479}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6830","nonce":3790896479},{"type":"v1","addr":"192.168.123.109:6831","nonce":3790896479}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6828","nonce":3790896479},{"type":"v1","addr":"192.168.123.109:6829","nonce":3790896479}]},"public_addr":"192.168.123.109:6825/3790896479","cluster_addr":"192.168.123.109:6827/3790896479","heartbeat_back_addr":"192.168.123.109:6831/3790896479","heartbeat_front_addr":"192.168.123.109:6829/3790896479","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:35.175154+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:35.231151+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:37.676669+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:36.715258+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:38.752713+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:39.447159+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:40.346150+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:40.261049+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.107:0/694361325":"2026-03-08T06:54:25.066469+0000","192.168.123.107:0/2795543388":"2026-03-08T06:54:25.066469+0000","192.168.123.107:0/1246254229":"2026-03-08T06:53:17.025312+0000","192.168.123.107:6801/2092210648":"2026-03-08T06:53:17.025312+0000","192.168.123.107:0/230558429":"2026-03-08T06:53:17.025312+0000","192.168.123.107:0/162808840":"2026-03-08T06:53:37.191091+0000","192.168.123.107:0/3758503808":"2026-03-08T06:53:17.025312+0000","192.168.123.107:0/3016678431":"2026-03-08T06:54:25.066469+0000","192.168.123.107:6801/2342336319":"2026-03-08T06:53:37.191091+0000","192.168.123.107:0/1760217480":"2026-03-08T06:53:37.191091+0000","192.168.123.107:6800/2342336319":"2026-03-08T06:53:37.191091+0000","192.168.123.107:0/2619217638":"2026-03-08T06:53:37.191091+0000","192.168.123.107:6800/3173836049":"2026-03-08T06:54:25.066469+0000","192.168.123.107:6800/2092210648":"2026-03-08T06:53:17.025312+0000","192.168.123.107:6801/3173836049":"2026-03-08T06:54:25.066469+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-07T07:55:50.864 INFO:tasks.cephadm.ceph_manager.ceph:[{'pool': 1, 'pool_name': '.mgr', 'create_time': '2026-03-07T06:55:41.135780+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'is_stretch_pool': False, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': "0'0", 'target_version': "0'0"}, 'last_change': '23', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {'mgr': {}}, 'read_balance': {'score_type': 'Fair distribution', 'score_acting': 7.889999866485596, 'score_stable': 7.889999866485596, 'optimal_score': 0.3799999952316284, 'raw_score_acting': 3, 'raw_score_stable': 3, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}}] 2026-03-07T07:55:50.865 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd pool get .mgr pg_num 2026-03-07T07:55:51.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:50 vm09 bash[21220]: audit 2026-03-07T06:55:50.781791+0000 mon.vm07 (mon.0) 611 : audit [DBG] from='client.? 192.168.123.107:0/1486471979' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T07:55:51.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:50 vm09 bash[21220]: audit 2026-03-07T06:55:50.781791+0000 mon.vm07 (mon.0) 611 : audit [DBG] from='client.? 192.168.123.107:0/1486471979' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T07:55:51.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:50 vm07 bash[17031]: audit 2026-03-07T06:55:50.781791+0000 mon.vm07 (mon.0) 611 : audit [DBG] from='client.? 192.168.123.107:0/1486471979' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T07:55:51.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:50 vm07 bash[17031]: audit 2026-03-07T06:55:50.781791+0000 mon.vm07 (mon.0) 611 : audit [DBG] from='client.? 192.168.123.107:0/1486471979' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T07:55:52.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:51 vm09 bash[21220]: cluster 2026-03-07T06:55:51.091669+0000 mgr.vm07.yrfcuj (mgr.14201) 96 : cluster [DBG] pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:55:52.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:51 vm09 bash[21220]: cluster 2026-03-07T06:55:51.091669+0000 mgr.vm07.yrfcuj (mgr.14201) 96 : cluster [DBG] pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:55:52.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:51 vm07 bash[17031]: cluster 2026-03-07T06:55:51.091669+0000 mgr.vm07.yrfcuj (mgr.14201) 96 : cluster [DBG] pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:55:52.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:51 vm07 bash[17031]: cluster 2026-03-07T06:55:51.091669+0000 mgr.vm07.yrfcuj (mgr.14201) 96 : cluster [DBG] pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:55:54.601 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:54 vm09 bash[21220]: cluster 2026-03-07T06:55:53.091942+0000 mgr.vm07.yrfcuj (mgr.14201) 97 : cluster [DBG] pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:55:54.601 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:54 vm09 bash[21220]: cluster 2026-03-07T06:55:53.091942+0000 mgr.vm07.yrfcuj (mgr.14201) 97 : cluster [DBG] pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:55:54.601 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:54 vm09 bash[21220]: audit 2026-03-07T06:55:53.730807+0000 mon.vm07 (mon.0) 612 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:54.601 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:54 vm09 bash[21220]: audit 2026-03-07T06:55:53.730807+0000 mon.vm07 (mon.0) 612 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:54.601 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:54 vm09 bash[21220]: audit 2026-03-07T06:55:53.738004+0000 mon.vm07 (mon.0) 613 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:54.601 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:54 vm09 bash[21220]: audit 2026-03-07T06:55:53.738004+0000 mon.vm07 (mon.0) 613 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:54.601 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:54 vm09 bash[21220]: audit 2026-03-07T06:55:53.739174+0000 mon.vm07 (mon.0) 614 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"}]: dispatch 2026-03-07T07:55:54.601 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:54 vm09 bash[21220]: audit 2026-03-07T06:55:53.739174+0000 mon.vm07 (mon.0) 614 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"}]: dispatch 2026-03-07T07:55:54.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:54 vm07 bash[17031]: cluster 2026-03-07T06:55:53.091942+0000 mgr.vm07.yrfcuj (mgr.14201) 97 : cluster [DBG] pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:55:54.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:54 vm07 bash[17031]: cluster 2026-03-07T06:55:53.091942+0000 mgr.vm07.yrfcuj (mgr.14201) 97 : cluster [DBG] pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:55:54.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:54 vm07 bash[17031]: audit 2026-03-07T06:55:53.730807+0000 mon.vm07 (mon.0) 612 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:54.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:54 vm07 bash[17031]: audit 2026-03-07T06:55:53.730807+0000 mon.vm07 (mon.0) 612 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:54.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:54 vm07 bash[17031]: audit 2026-03-07T06:55:53.738004+0000 mon.vm07 (mon.0) 613 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:54.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:54 vm07 bash[17031]: audit 2026-03-07T06:55:53.738004+0000 mon.vm07 (mon.0) 613 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:54.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:54 vm07 bash[17031]: audit 2026-03-07T06:55:53.739174+0000 mon.vm07 (mon.0) 614 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"}]: dispatch 2026-03-07T07:55:54.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:54 vm07 bash[17031]: audit 2026-03-07T06:55:53.739174+0000 mon.vm07 (mon.0) 614 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"}]: dispatch 2026-03-07T07:55:55.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:55 vm09 bash[21220]: cephadm 2026-03-07T06:55:53.725170+0000 mgr.vm07.yrfcuj (mgr.14201) 98 : cephadm [INF] Detected new or changed devices on vm07 2026-03-07T07:55:55.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:55 vm09 bash[21220]: cephadm 2026-03-07T06:55:53.725170+0000 mgr.vm07.yrfcuj (mgr.14201) 98 : cephadm [INF] Detected new or changed devices on vm07 2026-03-07T07:55:55.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:55 vm09 bash[21220]: audit 2026-03-07T06:55:54.870481+0000 mon.vm07 (mon.0) 615 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:55.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:55 vm09 bash[21220]: audit 2026-03-07T06:55:54.870481+0000 mon.vm07 (mon.0) 615 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:55.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:55 vm09 bash[21220]: audit 2026-03-07T06:55:54.874888+0000 mon.vm07 (mon.0) 616 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:55.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:55 vm09 bash[21220]: audit 2026-03-07T06:55:54.874888+0000 mon.vm07 (mon.0) 616 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:55.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:55 vm09 bash[21220]: audit 2026-03-07T06:55:54.875627+0000 mon.vm07 (mon.0) 617 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-07T07:55:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:55 vm09 bash[21220]: audit 2026-03-07T06:55:54.875627+0000 mon.vm07 (mon.0) 617 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-07T07:55:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:55 vm09 bash[21220]: audit 2026-03-07T06:55:54.876317+0000 mon.vm07 (mon.0) 618 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:55 vm09 bash[21220]: audit 2026-03-07T06:55:54.876317+0000 mon.vm07 (mon.0) 618 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:55 vm09 bash[21220]: audit 2026-03-07T06:55:54.876707+0000 mon.vm07 (mon.0) 619 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:55:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:55 vm09 bash[21220]: audit 2026-03-07T06:55:54.876707+0000 mon.vm07 (mon.0) 619 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:55:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:55 vm09 bash[21220]: audit 2026-03-07T06:55:54.879881+0000 mon.vm07 (mon.0) 620 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:55 vm09 bash[21220]: audit 2026-03-07T06:55:54.879881+0000 mon.vm07 (mon.0) 620 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:55 vm09 bash[21220]: audit 2026-03-07T06:55:54.881152+0000 mon.vm07 (mon.0) 621 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:55:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:55 vm09 bash[21220]: audit 2026-03-07T06:55:54.881152+0000 mon.vm07 (mon.0) 621 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:55:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:55 vm09 bash[21220]: audit 2026-03-07T06:55:55.125190+0000 mon.vm07 (mon.0) 622 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:55:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:55 vm09 bash[21220]: audit 2026-03-07T06:55:55.125190+0000 mon.vm07 (mon.0) 622 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:55:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:55 vm07 bash[17031]: cephadm 2026-03-07T06:55:53.725170+0000 mgr.vm07.yrfcuj (mgr.14201) 98 : cephadm [INF] Detected new or changed devices on vm07 2026-03-07T07:55:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:55 vm07 bash[17031]: cephadm 2026-03-07T06:55:53.725170+0000 mgr.vm07.yrfcuj (mgr.14201) 98 : cephadm [INF] Detected new or changed devices on vm07 2026-03-07T07:55:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:55 vm07 bash[17031]: audit 2026-03-07T06:55:54.870481+0000 mon.vm07 (mon.0) 615 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:55 vm07 bash[17031]: audit 2026-03-07T06:55:54.870481+0000 mon.vm07 (mon.0) 615 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:55 vm07 bash[17031]: audit 2026-03-07T06:55:54.874888+0000 mon.vm07 (mon.0) 616 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:55 vm07 bash[17031]: audit 2026-03-07T06:55:54.874888+0000 mon.vm07 (mon.0) 616 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:55 vm07 bash[17031]: audit 2026-03-07T06:55:54.875627+0000 mon.vm07 (mon.0) 617 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-07T07:55:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:55 vm07 bash[17031]: audit 2026-03-07T06:55:54.875627+0000 mon.vm07 (mon.0) 617 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-07T07:55:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:55 vm07 bash[17031]: audit 2026-03-07T06:55:54.876317+0000 mon.vm07 (mon.0) 618 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:55 vm07 bash[17031]: audit 2026-03-07T06:55:54.876317+0000 mon.vm07 (mon.0) 618 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:55:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:55 vm07 bash[17031]: audit 2026-03-07T06:55:54.876707+0000 mon.vm07 (mon.0) 619 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:55:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:55 vm07 bash[17031]: audit 2026-03-07T06:55:54.876707+0000 mon.vm07 (mon.0) 619 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:55:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:55 vm07 bash[17031]: audit 2026-03-07T06:55:54.879881+0000 mon.vm07 (mon.0) 620 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:55 vm07 bash[17031]: audit 2026-03-07T06:55:54.879881+0000 mon.vm07 (mon.0) 620 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:55:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:55 vm07 bash[17031]: audit 2026-03-07T06:55:54.881152+0000 mon.vm07 (mon.0) 621 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:55:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:55 vm07 bash[17031]: audit 2026-03-07T06:55:54.881152+0000 mon.vm07 (mon.0) 621 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:55:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:55 vm07 bash[17031]: audit 2026-03-07T06:55:55.125190+0000 mon.vm07 (mon.0) 622 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:55:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:55 vm07 bash[17031]: audit 2026-03-07T06:55:55.125190+0000 mon.vm07 (mon.0) 622 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:55:56.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:56 vm09 bash[21220]: cephadm 2026-03-07T06:55:54.861407+0000 mgr.vm07.yrfcuj (mgr.14201) 99 : cephadm [INF] Detected new or changed devices on vm09 2026-03-07T07:55:56.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:56 vm09 bash[21220]: cephadm 2026-03-07T06:55:54.861407+0000 mgr.vm07.yrfcuj (mgr.14201) 99 : cephadm [INF] Detected new or changed devices on vm09 2026-03-07T07:55:56.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:56 vm09 bash[21220]: cluster 2026-03-07T06:55:55.092176+0000 mgr.vm07.yrfcuj (mgr.14201) 100 : cluster [DBG] pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:55:56.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:56 vm09 bash[21220]: cluster 2026-03-07T06:55:55.092176+0000 mgr.vm07.yrfcuj (mgr.14201) 100 : cluster [DBG] pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:55:56.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:56 vm07 bash[17031]: cephadm 2026-03-07T06:55:54.861407+0000 mgr.vm07.yrfcuj (mgr.14201) 99 : cephadm [INF] Detected new or changed devices on vm09 2026-03-07T07:55:56.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:56 vm07 bash[17031]: cephadm 2026-03-07T06:55:54.861407+0000 mgr.vm07.yrfcuj (mgr.14201) 99 : cephadm [INF] Detected new or changed devices on vm09 2026-03-07T07:55:56.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:56 vm07 bash[17031]: cluster 2026-03-07T06:55:55.092176+0000 mgr.vm07.yrfcuj (mgr.14201) 100 : cluster [DBG] pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:55:56.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:56 vm07 bash[17031]: cluster 2026-03-07T06:55:55.092176+0000 mgr.vm07.yrfcuj (mgr.14201) 100 : cluster [DBG] pgmap v56: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:55:56.735 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:55:57.125 INFO:teuthology.orchestra.run.vm07.stdout:pg_num: 1 2026-03-07T07:55:57.190 INFO:tasks.cephadm:Setting up client nodes... 2026-03-07T07:55:57.190 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph auth get-or-create client.0 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-07T07:55:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:57 vm07 bash[17031]: audit 2026-03-07T06:55:57.125202+0000 mon.vm09 (mon.1) 18 : audit [DBG] from='client.? 192.168.123.107:0/309272297' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-07T07:55:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:57 vm07 bash[17031]: audit 2026-03-07T06:55:57.125202+0000 mon.vm09 (mon.1) 18 : audit [DBG] from='client.? 192.168.123.107:0/309272297' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-07T07:55:57.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:57 vm09 bash[21220]: audit 2026-03-07T06:55:57.125202+0000 mon.vm09 (mon.1) 18 : audit [DBG] from='client.? 192.168.123.107:0/309272297' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-07T07:55:57.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:57 vm09 bash[21220]: audit 2026-03-07T06:55:57.125202+0000 mon.vm09 (mon.1) 18 : audit [DBG] from='client.? 192.168.123.107:0/309272297' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-07T07:55:58.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:58 vm09 bash[21220]: cluster 2026-03-07T06:55:57.092467+0000 mgr.vm07.yrfcuj (mgr.14201) 101 : cluster [DBG] pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:55:58.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:55:58 vm09 bash[21220]: cluster 2026-03-07T06:55:57.092467+0000 mgr.vm07.yrfcuj (mgr.14201) 101 : cluster [DBG] pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:55:58.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:58 vm07 bash[17031]: cluster 2026-03-07T06:55:57.092467+0000 mgr.vm07.yrfcuj (mgr.14201) 101 : cluster [DBG] pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:55:58.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:55:58 vm07 bash[17031]: cluster 2026-03-07T06:55:57.092467+0000 mgr.vm07.yrfcuj (mgr.14201) 101 : cluster [DBG] pgmap v57: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:00.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:00 vm09 bash[21220]: cluster 2026-03-07T06:55:59.092757+0000 mgr.vm07.yrfcuj (mgr.14201) 102 : cluster [DBG] pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:00.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:00 vm09 bash[21220]: cluster 2026-03-07T06:55:59.092757+0000 mgr.vm07.yrfcuj (mgr.14201) 102 : cluster [DBG] pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:00.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:00 vm07 bash[17031]: cluster 2026-03-07T06:55:59.092757+0000 mgr.vm07.yrfcuj (mgr.14201) 102 : cluster [DBG] pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:00.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:00 vm07 bash[17031]: cluster 2026-03-07T06:55:59.092757+0000 mgr.vm07.yrfcuj (mgr.14201) 102 : cluster [DBG] pgmap v58: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:01.985 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:02.384 INFO:teuthology.orchestra.run.vm07.stdout:[client.0] 2026-03-07T07:56:02.384 INFO:teuthology.orchestra.run.vm07.stdout: key = AQACzKtp+71gFhAA2EFdgHpqZyg2FlYiC5Z/IQ== 2026-03-07T07:56:02.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:02 vm07 bash[17031]: cluster 2026-03-07T06:56:01.093027+0000 mgr.vm07.yrfcuj (mgr.14201) 103 : cluster [DBG] pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:02.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:02 vm07 bash[17031]: cluster 2026-03-07T06:56:01.093027+0000 mgr.vm07.yrfcuj (mgr.14201) 103 : cluster [DBG] pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:02.446 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-07T07:56:02.446 DEBUG:teuthology.orchestra.run.vm07:> sudo dd of=/etc/ceph/ceph.client.0.keyring 2026-03-07T07:56:02.446 DEBUG:teuthology.orchestra.run.vm07:> sudo chmod 0644 /etc/ceph/ceph.client.0.keyring 2026-03-07T07:56:02.457 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph auth get-or-create client.1 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-07T07:56:02.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:02 vm09 bash[21220]: cluster 2026-03-07T06:56:01.093027+0000 mgr.vm07.yrfcuj (mgr.14201) 103 : cluster [DBG] pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:02.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:02 vm09 bash[21220]: cluster 2026-03-07T06:56:01.093027+0000 mgr.vm07.yrfcuj (mgr.14201) 103 : cluster [DBG] pgmap v59: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:03.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:03 vm09 bash[21220]: audit 2026-03-07T06:56:02.375320+0000 mon.vm07 (mon.0) 623 : audit [INF] from='client.? 192.168.123.107:0/3585917220' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-07T07:56:03.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:03 vm09 bash[21220]: audit 2026-03-07T06:56:02.375320+0000 mon.vm07 (mon.0) 623 : audit [INF] from='client.? 192.168.123.107:0/3585917220' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-07T07:56:03.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:03 vm09 bash[21220]: audit 2026-03-07T06:56:02.377905+0000 mon.vm07 (mon.0) 624 : audit [INF] from='client.? 192.168.123.107:0/3585917220' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-07T07:56:03.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:03 vm09 bash[21220]: audit 2026-03-07T06:56:02.377905+0000 mon.vm07 (mon.0) 624 : audit [INF] from='client.? 192.168.123.107:0/3585917220' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-07T07:56:03.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:03 vm07 bash[17031]: audit 2026-03-07T06:56:02.375320+0000 mon.vm07 (mon.0) 623 : audit [INF] from='client.? 192.168.123.107:0/3585917220' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-07T07:56:03.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:03 vm07 bash[17031]: audit 2026-03-07T06:56:02.375320+0000 mon.vm07 (mon.0) 623 : audit [INF] from='client.? 192.168.123.107:0/3585917220' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-07T07:56:03.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:03 vm07 bash[17031]: audit 2026-03-07T06:56:02.377905+0000 mon.vm07 (mon.0) 624 : audit [INF] from='client.? 192.168.123.107:0/3585917220' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-07T07:56:03.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:03 vm07 bash[17031]: audit 2026-03-07T06:56:02.377905+0000 mon.vm07 (mon.0) 624 : audit [INF] from='client.? 192.168.123.107:0/3585917220' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-07T07:56:04.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:04 vm07 bash[17031]: cluster 2026-03-07T06:56:03.093312+0000 mgr.vm07.yrfcuj (mgr.14201) 104 : cluster [DBG] pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:04.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:04 vm07 bash[17031]: cluster 2026-03-07T06:56:03.093312+0000 mgr.vm07.yrfcuj (mgr.14201) 104 : cluster [DBG] pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:04.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:04 vm09 bash[21220]: cluster 2026-03-07T06:56:03.093312+0000 mgr.vm07.yrfcuj (mgr.14201) 104 : cluster [DBG] pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:04.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:04 vm09 bash[21220]: cluster 2026-03-07T06:56:03.093312+0000 mgr.vm07.yrfcuj (mgr.14201) 104 : cluster [DBG] pgmap v60: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:06.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:06 vm07 bash[17031]: cluster 2026-03-07T06:56:05.093585+0000 mgr.vm07.yrfcuj (mgr.14201) 105 : cluster [DBG] pgmap v61: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:06.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:06 vm07 bash[17031]: cluster 2026-03-07T06:56:05.093585+0000 mgr.vm07.yrfcuj (mgr.14201) 105 : cluster [DBG] pgmap v61: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:06.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:06 vm09 bash[21220]: cluster 2026-03-07T06:56:05.093585+0000 mgr.vm07.yrfcuj (mgr.14201) 105 : cluster [DBG] pgmap v61: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:06.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:06 vm09 bash[21220]: cluster 2026-03-07T06:56:05.093585+0000 mgr.vm07.yrfcuj (mgr.14201) 105 : cluster [DBG] pgmap v61: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:07.224 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm09/config 2026-03-07T07:56:07.628 INFO:teuthology.orchestra.run.vm09.stdout:[client.1] 2026-03-07T07:56:07.628 INFO:teuthology.orchestra.run.vm09.stdout: key = AQAHzKtpEqfZJBAA/b0CbQk1iWSgFnIKnL7GYw== 2026-03-07T07:56:07.687 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-07T07:56:07.687 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/ceph.client.1.keyring 2026-03-07T07:56:07.687 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod 0644 /etc/ceph/ceph.client.1.keyring 2026-03-07T07:56:07.698 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-07T07:56:07.698 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-07T07:56:07.698 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph mgr dump --format=json 2026-03-07T07:56:08.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:08 vm07 bash[17031]: cluster 2026-03-07T06:56:07.093822+0000 mgr.vm07.yrfcuj (mgr.14201) 106 : cluster [DBG] pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:08.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:08 vm07 bash[17031]: cluster 2026-03-07T06:56:07.093822+0000 mgr.vm07.yrfcuj (mgr.14201) 106 : cluster [DBG] pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:08.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:08 vm07 bash[17031]: audit 2026-03-07T06:56:07.618142+0000 mon.vm07 (mon.0) 625 : audit [INF] from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-07T07:56:08.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:08 vm07 bash[17031]: audit 2026-03-07T06:56:07.618142+0000 mon.vm07 (mon.0) 625 : audit [INF] from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-07T07:56:08.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:08 vm07 bash[17031]: audit 2026-03-07T06:56:07.620324+0000 mon.vm07 (mon.0) 626 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-07T07:56:08.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:08 vm07 bash[17031]: audit 2026-03-07T06:56:07.620324+0000 mon.vm07 (mon.0) 626 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-07T07:56:08.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:08 vm07 bash[17031]: audit 2026-03-07T06:56:07.622861+0000 mon.vm09 (mon.1) 19 : audit [INF] from='client.? 192.168.123.109:0/2732680598' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-07T07:56:08.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:08 vm07 bash[17031]: audit 2026-03-07T06:56:07.622861+0000 mon.vm09 (mon.1) 19 : audit [INF] from='client.? 192.168.123.109:0/2732680598' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-07T07:56:08.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:08 vm09 bash[21220]: cluster 2026-03-07T06:56:07.093822+0000 mgr.vm07.yrfcuj (mgr.14201) 106 : cluster [DBG] pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:08.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:08 vm09 bash[21220]: cluster 2026-03-07T06:56:07.093822+0000 mgr.vm07.yrfcuj (mgr.14201) 106 : cluster [DBG] pgmap v62: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:08.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:08 vm09 bash[21220]: audit 2026-03-07T06:56:07.618142+0000 mon.vm07 (mon.0) 625 : audit [INF] from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-07T07:56:08.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:08 vm09 bash[21220]: audit 2026-03-07T06:56:07.618142+0000 mon.vm07 (mon.0) 625 : audit [INF] from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-07T07:56:08.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:08 vm09 bash[21220]: audit 2026-03-07T06:56:07.620324+0000 mon.vm07 (mon.0) 626 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-07T07:56:08.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:08 vm09 bash[21220]: audit 2026-03-07T06:56:07.620324+0000 mon.vm07 (mon.0) 626 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-07T07:56:08.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:08 vm09 bash[21220]: audit 2026-03-07T06:56:07.622861+0000 mon.vm09 (mon.1) 19 : audit [INF] from='client.? 192.168.123.109:0/2732680598' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-07T07:56:08.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:08 vm09 bash[21220]: audit 2026-03-07T06:56:07.622861+0000 mon.vm09 (mon.1) 19 : audit [INF] from='client.? 192.168.123.109:0/2732680598' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-07T07:56:10.473 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:10 vm07 bash[17031]: cluster 2026-03-07T06:56:09.094163+0000 mgr.vm07.yrfcuj (mgr.14201) 107 : cluster [DBG] pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:10.473 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:10 vm07 bash[17031]: cluster 2026-03-07T06:56:09.094163+0000 mgr.vm07.yrfcuj (mgr.14201) 107 : cluster [DBG] pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:10.473 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:10 vm07 bash[17031]: audit 2026-03-07T06:56:10.125599+0000 mon.vm07 (mon.0) 627 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:56:10.473 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:10 vm07 bash[17031]: audit 2026-03-07T06:56:10.125599+0000 mon.vm07 (mon.0) 627 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:56:10.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:10 vm09 bash[21220]: cluster 2026-03-07T06:56:09.094163+0000 mgr.vm07.yrfcuj (mgr.14201) 107 : cluster [DBG] pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:10.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:10 vm09 bash[21220]: cluster 2026-03-07T06:56:09.094163+0000 mgr.vm07.yrfcuj (mgr.14201) 107 : cluster [DBG] pgmap v63: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:10.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:10 vm09 bash[21220]: audit 2026-03-07T06:56:10.125599+0000 mon.vm07 (mon.0) 627 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:56:10.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:10 vm09 bash[21220]: audit 2026-03-07T06:56:10.125599+0000 mon.vm07 (mon.0) 627 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:56:12.496 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:12.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:12 vm09 bash[21220]: cluster 2026-03-07T06:56:11.094427+0000 mgr.vm07.yrfcuj (mgr.14201) 108 : cluster [DBG] pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:12.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:12 vm09 bash[21220]: cluster 2026-03-07T06:56:11.094427+0000 mgr.vm07.yrfcuj (mgr.14201) 108 : cluster [DBG] pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:12.905 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:56:12.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:12 vm07 bash[17031]: cluster 2026-03-07T06:56:11.094427+0000 mgr.vm07.yrfcuj (mgr.14201) 108 : cluster [DBG] pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:12.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:12 vm07 bash[17031]: cluster 2026-03-07T06:56:11.094427+0000 mgr.vm07.yrfcuj (mgr.14201) 108 : cluster [DBG] pgmap v64: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:12.979 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":19,"flags":0,"active_gid":14201,"active_name":"vm07.yrfcuj","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6800","nonce":3942096239},{"type":"v1","addr":"192.168.123.107:6801","nonce":3942096239}]},"active_addr":"192.168.123.107:6801/3942096239","active_change":"2026-03-07T06:54:25.066719+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[{"gid":14220,"name":"vm09.eqznpw","mgr_features":4540701547738038271,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","prometheus","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.107:8443/","prometheus":"http://192.168.123.107:9283/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":5,"active_clients":[{"name":"devicehealth","addrvec":[{"type":"v2","addr":"192.168.123.107:0","nonce":3034081729}]},{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.107:0","nonce":3038943207}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.107:0","nonce":3207151706}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.107:0","nonce":1689887044}]}]} 2026-03-07T07:56:12.981 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-07T07:56:12.982 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-07T07:56:12.982 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd dump --format=json 2026-03-07T07:56:13.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:13 vm09 bash[21220]: audit 2026-03-07T06:56:12.897255+0000 mon.vm07 (mon.0) 628 : audit [DBG] from='client.? 192.168.123.107:0/1840898328' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-07T07:56:13.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:13 vm09 bash[21220]: audit 2026-03-07T06:56:12.897255+0000 mon.vm07 (mon.0) 628 : audit [DBG] from='client.? 192.168.123.107:0/1840898328' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-07T07:56:13.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:13 vm07 bash[17031]: audit 2026-03-07T06:56:12.897255+0000 mon.vm07 (mon.0) 628 : audit [DBG] from='client.? 192.168.123.107:0/1840898328' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-07T07:56:13.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:13 vm07 bash[17031]: audit 2026-03-07T06:56:12.897255+0000 mon.vm07 (mon.0) 628 : audit [DBG] from='client.? 192.168.123.107:0/1840898328' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-07T07:56:14.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:14 vm09 bash[21220]: cluster 2026-03-07T06:56:13.094750+0000 mgr.vm07.yrfcuj (mgr.14201) 109 : cluster [DBG] pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:14.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:14 vm09 bash[21220]: cluster 2026-03-07T06:56:13.094750+0000 mgr.vm07.yrfcuj (mgr.14201) 109 : cluster [DBG] pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:14.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:14 vm07 bash[17031]: cluster 2026-03-07T06:56:13.094750+0000 mgr.vm07.yrfcuj (mgr.14201) 109 : cluster [DBG] pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:14.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:14 vm07 bash[17031]: cluster 2026-03-07T06:56:13.094750+0000 mgr.vm07.yrfcuj (mgr.14201) 109 : cluster [DBG] pgmap v65: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:16.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:16 vm09 bash[21220]: cluster 2026-03-07T06:56:15.095003+0000 mgr.vm07.yrfcuj (mgr.14201) 110 : cluster [DBG] pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:16.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:16 vm09 bash[21220]: cluster 2026-03-07T06:56:15.095003+0000 mgr.vm07.yrfcuj (mgr.14201) 110 : cluster [DBG] pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:16.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:16 vm07 bash[17031]: cluster 2026-03-07T06:56:15.095003+0000 mgr.vm07.yrfcuj (mgr.14201) 110 : cluster [DBG] pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:16.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:16 vm07 bash[17031]: cluster 2026-03-07T06:56:15.095003+0000 mgr.vm07.yrfcuj (mgr.14201) 110 : cluster [DBG] pgmap v66: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:17.767 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:18.128 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:56:18.129 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":23,"fsid":"312fdbc4-19f2-11f1-81d0-4bbd10a1e012","created":"2026-03-07T06:52:52.038690+0000","modified":"2026-03-07T06:55:44.240571+0000","last_up_change":"2026-03-07T06:55:41.609166+0000","last_in_change":"2026-03-07T06:55:18.145593+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":9,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-07T06:55:41.135780+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"23","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"e6429ca8-de99-4811-a21e-2292aedf25ef","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":635544560},{"type":"v1","addr":"192.168.123.109:6801","nonce":635544560}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":635544560},{"type":"v1","addr":"192.168.123.109:6803","nonce":635544560}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":635544560},{"type":"v1","addr":"192.168.123.109:6807","nonce":635544560}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":635544560},{"type":"v1","addr":"192.168.123.109:6805","nonce":635544560}]},"public_addr":"192.168.123.109:6801/635544560","cluster_addr":"192.168.123.109:6803/635544560","heartbeat_back_addr":"192.168.123.109:6807/635544560","heartbeat_front_addr":"192.168.123.109:6805/635544560","state":["exists","up"]},{"osd":1,"uuid":"56cba12f-96c0-445c-be10-c84fa665caa4","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6802","nonce":1359088522},{"type":"v1","addr":"192.168.123.107:6803","nonce":1359088522}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6804","nonce":1359088522},{"type":"v1","addr":"192.168.123.107:6805","nonce":1359088522}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6808","nonce":1359088522},{"type":"v1","addr":"192.168.123.107:6809","nonce":1359088522}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6806","nonce":1359088522},{"type":"v1","addr":"192.168.123.107:6807","nonce":1359088522}]},"public_addr":"192.168.123.107:6803/1359088522","cluster_addr":"192.168.123.107:6805/1359088522","heartbeat_back_addr":"192.168.123.107:6809/1359088522","heartbeat_front_addr":"192.168.123.107:6807/1359088522","state":["exists","up"]},{"osd":2,"uuid":"2a313541-d79f-4e9b-8bb5-038ec51951db","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6810","nonce":158678589},{"type":"v1","addr":"192.168.123.107:6811","nonce":158678589}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6812","nonce":158678589},{"type":"v1","addr":"192.168.123.107:6813","nonce":158678589}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6816","nonce":158678589},{"type":"v1","addr":"192.168.123.107:6817","nonce":158678589}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6814","nonce":158678589},{"type":"v1","addr":"192.168.123.107:6815","nonce":158678589}]},"public_addr":"192.168.123.107:6811/158678589","cluster_addr":"192.168.123.107:6813/158678589","heartbeat_back_addr":"192.168.123.107:6817/158678589","heartbeat_front_addr":"192.168.123.107:6815/158678589","state":["exists","up"]},{"osd":3,"uuid":"a583cddd-35d0-41d0-83c4-877b3c089e8b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":19,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6808","nonce":1224712665},{"type":"v1","addr":"192.168.123.109:6809","nonce":1224712665}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6810","nonce":1224712665},{"type":"v1","addr":"192.168.123.109:6811","nonce":1224712665}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6814","nonce":1224712665},{"type":"v1","addr":"192.168.123.109:6815","nonce":1224712665}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6812","nonce":1224712665},{"type":"v1","addr":"192.168.123.109:6813","nonce":1224712665}]},"public_addr":"192.168.123.109:6809/1224712665","cluster_addr":"192.168.123.109:6811/1224712665","heartbeat_back_addr":"192.168.123.109:6815/1224712665","heartbeat_front_addr":"192.168.123.109:6813/1224712665","state":["exists","up"]},{"osd":4,"uuid":"2f651ee5-5a7b-4d0c-82f4-907f0a8beddc","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6816","nonce":1602816023},{"type":"v1","addr":"192.168.123.109:6817","nonce":1602816023}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6818","nonce":1602816023},{"type":"v1","addr":"192.168.123.109:6819","nonce":1602816023}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6822","nonce":1602816023},{"type":"v1","addr":"192.168.123.109:6823","nonce":1602816023}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6820","nonce":1602816023},{"type":"v1","addr":"192.168.123.109:6821","nonce":1602816023}]},"public_addr":"192.168.123.109:6817/1602816023","cluster_addr":"192.168.123.109:6819/1602816023","heartbeat_back_addr":"192.168.123.109:6823/1602816023","heartbeat_front_addr":"192.168.123.109:6821/1602816023","state":["exists","up"]},{"osd":5,"uuid":"5874e093-78e8-4710-8cf9-d27b73d59b9b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6818","nonce":212397676},{"type":"v1","addr":"192.168.123.107:6819","nonce":212397676}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6820","nonce":212397676},{"type":"v1","addr":"192.168.123.107:6821","nonce":212397676}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6824","nonce":212397676},{"type":"v1","addr":"192.168.123.107:6825","nonce":212397676}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6822","nonce":212397676},{"type":"v1","addr":"192.168.123.107:6823","nonce":212397676}]},"public_addr":"192.168.123.107:6819/212397676","cluster_addr":"192.168.123.107:6821/212397676","heartbeat_back_addr":"192.168.123.107:6825/212397676","heartbeat_front_addr":"192.168.123.107:6823/212397676","state":["exists","up"]},{"osd":6,"uuid":"85f57ef2-7375-4150-b632-6d08ce1f021e","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6826","nonce":2603438728},{"type":"v1","addr":"192.168.123.107:6827","nonce":2603438728}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6828","nonce":2603438728},{"type":"v1","addr":"192.168.123.107:6829","nonce":2603438728}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6832","nonce":2603438728},{"type":"v1","addr":"192.168.123.107:6833","nonce":2603438728}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6830","nonce":2603438728},{"type":"v1","addr":"192.168.123.107:6831","nonce":2603438728}]},"public_addr":"192.168.123.107:6827/2603438728","cluster_addr":"192.168.123.107:6829/2603438728","heartbeat_back_addr":"192.168.123.107:6833/2603438728","heartbeat_front_addr":"192.168.123.107:6831/2603438728","state":["exists","up"]},{"osd":7,"uuid":"6ea2bf16-f584-49a8-ba45-a4b9ba5043ce","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":21,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6824","nonce":3790896479},{"type":"v1","addr":"192.168.123.109:6825","nonce":3790896479}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6826","nonce":3790896479},{"type":"v1","addr":"192.168.123.109:6827","nonce":3790896479}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6830","nonce":3790896479},{"type":"v1","addr":"192.168.123.109:6831","nonce":3790896479}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6828","nonce":3790896479},{"type":"v1","addr":"192.168.123.109:6829","nonce":3790896479}]},"public_addr":"192.168.123.109:6825/3790896479","cluster_addr":"192.168.123.109:6827/3790896479","heartbeat_back_addr":"192.168.123.109:6831/3790896479","heartbeat_front_addr":"192.168.123.109:6829/3790896479","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:35.175154+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:35.231151+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:37.676669+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:36.715258+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:38.752713+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:39.447159+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:40.346150+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:40.261049+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.107:0/694361325":"2026-03-08T06:54:25.066469+0000","192.168.123.107:0/2795543388":"2026-03-08T06:54:25.066469+0000","192.168.123.107:0/1246254229":"2026-03-08T06:53:17.025312+0000","192.168.123.107:6801/2092210648":"2026-03-08T06:53:17.025312+0000","192.168.123.107:0/230558429":"2026-03-08T06:53:17.025312+0000","192.168.123.107:0/162808840":"2026-03-08T06:53:37.191091+0000","192.168.123.107:0/3758503808":"2026-03-08T06:53:17.025312+0000","192.168.123.107:0/3016678431":"2026-03-08T06:54:25.066469+0000","192.168.123.107:6801/2342336319":"2026-03-08T06:53:37.191091+0000","192.168.123.107:0/1760217480":"2026-03-08T06:53:37.191091+0000","192.168.123.107:6800/2342336319":"2026-03-08T06:53:37.191091+0000","192.168.123.107:0/2619217638":"2026-03-08T06:53:37.191091+0000","192.168.123.107:6800/3173836049":"2026-03-08T06:54:25.066469+0000","192.168.123.107:6800/2092210648":"2026-03-08T06:53:17.025312+0000","192.168.123.107:6801/3173836049":"2026-03-08T06:54:25.066469+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-07T07:56:18.192 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-07T07:56:18.192 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd dump --format=json 2026-03-07T07:56:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:18 vm07 bash[17031]: cluster 2026-03-07T06:56:17.095276+0000 mgr.vm07.yrfcuj (mgr.14201) 111 : cluster [DBG] pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:18 vm07 bash[17031]: cluster 2026-03-07T06:56:17.095276+0000 mgr.vm07.yrfcuj (mgr.14201) 111 : cluster [DBG] pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:18 vm07 bash[17031]: audit 2026-03-07T06:56:18.122795+0000 mon.vm07 (mon.0) 629 : audit [DBG] from='client.? 192.168.123.107:0/2527398954' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T07:56:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:18 vm07 bash[17031]: audit 2026-03-07T06:56:18.122795+0000 mon.vm07 (mon.0) 629 : audit [DBG] from='client.? 192.168.123.107:0/2527398954' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T07:56:18.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:18 vm09 bash[21220]: cluster 2026-03-07T06:56:17.095276+0000 mgr.vm07.yrfcuj (mgr.14201) 111 : cluster [DBG] pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:18.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:18 vm09 bash[21220]: cluster 2026-03-07T06:56:17.095276+0000 mgr.vm07.yrfcuj (mgr.14201) 111 : cluster [DBG] pgmap v67: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:18.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:18 vm09 bash[21220]: audit 2026-03-07T06:56:18.122795+0000 mon.vm07 (mon.0) 629 : audit [DBG] from='client.? 192.168.123.107:0/2527398954' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T07:56:18.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:18 vm09 bash[21220]: audit 2026-03-07T06:56:18.122795+0000 mon.vm07 (mon.0) 629 : audit [DBG] from='client.? 192.168.123.107:0/2527398954' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T07:56:20.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:20 vm09 bash[21220]: cluster 2026-03-07T06:56:19.095575+0000 mgr.vm07.yrfcuj (mgr.14201) 112 : cluster [DBG] pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:20.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:20 vm09 bash[21220]: cluster 2026-03-07T06:56:19.095575+0000 mgr.vm07.yrfcuj (mgr.14201) 112 : cluster [DBG] pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:20.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:20 vm07 bash[17031]: cluster 2026-03-07T06:56:19.095575+0000 mgr.vm07.yrfcuj (mgr.14201) 112 : cluster [DBG] pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:20.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:20 vm07 bash[17031]: cluster 2026-03-07T06:56:19.095575+0000 mgr.vm07.yrfcuj (mgr.14201) 112 : cluster [DBG] pgmap v68: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:22.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:22 vm09 bash[21220]: cluster 2026-03-07T06:56:21.095839+0000 mgr.vm07.yrfcuj (mgr.14201) 113 : cluster [DBG] pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:22.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:22 vm09 bash[21220]: cluster 2026-03-07T06:56:21.095839+0000 mgr.vm07.yrfcuj (mgr.14201) 113 : cluster [DBG] pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:22.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:22 vm07 bash[17031]: cluster 2026-03-07T06:56:21.095839+0000 mgr.vm07.yrfcuj (mgr.14201) 113 : cluster [DBG] pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:22.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:22 vm07 bash[17031]: cluster 2026-03-07T06:56:21.095839+0000 mgr.vm07.yrfcuj (mgr.14201) 113 : cluster [DBG] pgmap v69: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:23.004 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:23.361 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:56:23.361 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":23,"fsid":"312fdbc4-19f2-11f1-81d0-4bbd10a1e012","created":"2026-03-07T06:52:52.038690+0000","modified":"2026-03-07T06:55:44.240571+0000","last_up_change":"2026-03-07T06:55:41.609166+0000","last_in_change":"2026-03-07T06:55:18.145593+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":9,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-07T06:55:41.135780+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"23","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"e6429ca8-de99-4811-a21e-2292aedf25ef","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":635544560},{"type":"v1","addr":"192.168.123.109:6801","nonce":635544560}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":635544560},{"type":"v1","addr":"192.168.123.109:6803","nonce":635544560}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":635544560},{"type":"v1","addr":"192.168.123.109:6807","nonce":635544560}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":635544560},{"type":"v1","addr":"192.168.123.109:6805","nonce":635544560}]},"public_addr":"192.168.123.109:6801/635544560","cluster_addr":"192.168.123.109:6803/635544560","heartbeat_back_addr":"192.168.123.109:6807/635544560","heartbeat_front_addr":"192.168.123.109:6805/635544560","state":["exists","up"]},{"osd":1,"uuid":"56cba12f-96c0-445c-be10-c84fa665caa4","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":16,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6802","nonce":1359088522},{"type":"v1","addr":"192.168.123.107:6803","nonce":1359088522}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6804","nonce":1359088522},{"type":"v1","addr":"192.168.123.107:6805","nonce":1359088522}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6808","nonce":1359088522},{"type":"v1","addr":"192.168.123.107:6809","nonce":1359088522}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6806","nonce":1359088522},{"type":"v1","addr":"192.168.123.107:6807","nonce":1359088522}]},"public_addr":"192.168.123.107:6803/1359088522","cluster_addr":"192.168.123.107:6805/1359088522","heartbeat_back_addr":"192.168.123.107:6809/1359088522","heartbeat_front_addr":"192.168.123.107:6807/1359088522","state":["exists","up"]},{"osd":2,"uuid":"2a313541-d79f-4e9b-8bb5-038ec51951db","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6810","nonce":158678589},{"type":"v1","addr":"192.168.123.107:6811","nonce":158678589}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6812","nonce":158678589},{"type":"v1","addr":"192.168.123.107:6813","nonce":158678589}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6816","nonce":158678589},{"type":"v1","addr":"192.168.123.107:6817","nonce":158678589}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6814","nonce":158678589},{"type":"v1","addr":"192.168.123.107:6815","nonce":158678589}]},"public_addr":"192.168.123.107:6811/158678589","cluster_addr":"192.168.123.107:6813/158678589","heartbeat_back_addr":"192.168.123.107:6817/158678589","heartbeat_front_addr":"192.168.123.107:6815/158678589","state":["exists","up"]},{"osd":3,"uuid":"a583cddd-35d0-41d0-83c4-877b3c089e8b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":19,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6808","nonce":1224712665},{"type":"v1","addr":"192.168.123.109:6809","nonce":1224712665}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6810","nonce":1224712665},{"type":"v1","addr":"192.168.123.109:6811","nonce":1224712665}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6814","nonce":1224712665},{"type":"v1","addr":"192.168.123.109:6815","nonce":1224712665}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6812","nonce":1224712665},{"type":"v1","addr":"192.168.123.109:6813","nonce":1224712665}]},"public_addr":"192.168.123.109:6809/1224712665","cluster_addr":"192.168.123.109:6811/1224712665","heartbeat_back_addr":"192.168.123.109:6815/1224712665","heartbeat_front_addr":"192.168.123.109:6813/1224712665","state":["exists","up"]},{"osd":4,"uuid":"2f651ee5-5a7b-4d0c-82f4-907f0a8beddc","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6816","nonce":1602816023},{"type":"v1","addr":"192.168.123.109:6817","nonce":1602816023}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6818","nonce":1602816023},{"type":"v1","addr":"192.168.123.109:6819","nonce":1602816023}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6822","nonce":1602816023},{"type":"v1","addr":"192.168.123.109:6823","nonce":1602816023}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6820","nonce":1602816023},{"type":"v1","addr":"192.168.123.109:6821","nonce":1602816023}]},"public_addr":"192.168.123.109:6817/1602816023","cluster_addr":"192.168.123.109:6819/1602816023","heartbeat_back_addr":"192.168.123.109:6823/1602816023","heartbeat_front_addr":"192.168.123.109:6821/1602816023","state":["exists","up"]},{"osd":5,"uuid":"5874e093-78e8-4710-8cf9-d27b73d59b9b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6818","nonce":212397676},{"type":"v1","addr":"192.168.123.107:6819","nonce":212397676}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6820","nonce":212397676},{"type":"v1","addr":"192.168.123.107:6821","nonce":212397676}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6824","nonce":212397676},{"type":"v1","addr":"192.168.123.107:6825","nonce":212397676}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6822","nonce":212397676},{"type":"v1","addr":"192.168.123.107:6823","nonce":212397676}]},"public_addr":"192.168.123.107:6819/212397676","cluster_addr":"192.168.123.107:6821/212397676","heartbeat_back_addr":"192.168.123.107:6825/212397676","heartbeat_front_addr":"192.168.123.107:6823/212397676","state":["exists","up"]},{"osd":6,"uuid":"85f57ef2-7375-4150-b632-6d08ce1f021e","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6826","nonce":2603438728},{"type":"v1","addr":"192.168.123.107:6827","nonce":2603438728}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6828","nonce":2603438728},{"type":"v1","addr":"192.168.123.107:6829","nonce":2603438728}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6832","nonce":2603438728},{"type":"v1","addr":"192.168.123.107:6833","nonce":2603438728}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6830","nonce":2603438728},{"type":"v1","addr":"192.168.123.107:6831","nonce":2603438728}]},"public_addr":"192.168.123.107:6827/2603438728","cluster_addr":"192.168.123.107:6829/2603438728","heartbeat_back_addr":"192.168.123.107:6833/2603438728","heartbeat_front_addr":"192.168.123.107:6831/2603438728","state":["exists","up"]},{"osd":7,"uuid":"6ea2bf16-f584-49a8-ba45-a4b9ba5043ce","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":21,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6824","nonce":3790896479},{"type":"v1","addr":"192.168.123.109:6825","nonce":3790896479}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6826","nonce":3790896479},{"type":"v1","addr":"192.168.123.109:6827","nonce":3790896479}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6830","nonce":3790896479},{"type":"v1","addr":"192.168.123.109:6831","nonce":3790896479}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6828","nonce":3790896479},{"type":"v1","addr":"192.168.123.109:6829","nonce":3790896479}]},"public_addr":"192.168.123.109:6825/3790896479","cluster_addr":"192.168.123.109:6827/3790896479","heartbeat_back_addr":"192.168.123.109:6831/3790896479","heartbeat_front_addr":"192.168.123.109:6829/3790896479","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:35.175154+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:35.231151+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:37.676669+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:36.715258+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:38.752713+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:39.447159+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:40.346150+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-07T06:55:40.261049+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.107:0/694361325":"2026-03-08T06:54:25.066469+0000","192.168.123.107:0/2795543388":"2026-03-08T06:54:25.066469+0000","192.168.123.107:0/1246254229":"2026-03-08T06:53:17.025312+0000","192.168.123.107:6801/2092210648":"2026-03-08T06:53:17.025312+0000","192.168.123.107:0/230558429":"2026-03-08T06:53:17.025312+0000","192.168.123.107:0/162808840":"2026-03-08T06:53:37.191091+0000","192.168.123.107:0/3758503808":"2026-03-08T06:53:17.025312+0000","192.168.123.107:0/3016678431":"2026-03-08T06:54:25.066469+0000","192.168.123.107:6801/2342336319":"2026-03-08T06:53:37.191091+0000","192.168.123.107:0/1760217480":"2026-03-08T06:53:37.191091+0000","192.168.123.107:6800/2342336319":"2026-03-08T06:53:37.191091+0000","192.168.123.107:0/2619217638":"2026-03-08T06:53:37.191091+0000","192.168.123.107:6800/3173836049":"2026-03-08T06:54:25.066469+0000","192.168.123.107:6800/2092210648":"2026-03-08T06:53:17.025312+0000","192.168.123.107:6801/3173836049":"2026-03-08T06:54:25.066469+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-07T07:56:23.419 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph tell osd.0 flush_pg_stats 2026-03-07T07:56:23.419 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph tell osd.1 flush_pg_stats 2026-03-07T07:56:23.419 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph tell osd.2 flush_pg_stats 2026-03-07T07:56:23.420 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph tell osd.3 flush_pg_stats 2026-03-07T07:56:23.420 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph tell osd.4 flush_pg_stats 2026-03-07T07:56:23.420 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph tell osd.5 flush_pg_stats 2026-03-07T07:56:23.420 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph tell osd.6 flush_pg_stats 2026-03-07T07:56:23.420 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph tell osd.7 flush_pg_stats 2026-03-07T07:56:23.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:23 vm07 bash[17031]: audit 2026-03-07T06:56:23.355819+0000 mon.vm07 (mon.0) 630 : audit [DBG] from='client.? 192.168.123.107:0/3214172770' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T07:56:23.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:23 vm07 bash[17031]: audit 2026-03-07T06:56:23.355819+0000 mon.vm07 (mon.0) 630 : audit [DBG] from='client.? 192.168.123.107:0/3214172770' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T07:56:23.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:23 vm09 bash[21220]: audit 2026-03-07T06:56:23.355819+0000 mon.vm07 (mon.0) 630 : audit [DBG] from='client.? 192.168.123.107:0/3214172770' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T07:56:23.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:23 vm09 bash[21220]: audit 2026-03-07T06:56:23.355819+0000 mon.vm07 (mon.0) 630 : audit [DBG] from='client.? 192.168.123.107:0/3214172770' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-07T07:56:24.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:24 vm09 bash[21220]: cluster 2026-03-07T06:56:23.096115+0000 mgr.vm07.yrfcuj (mgr.14201) 114 : cluster [DBG] pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:24.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:24 vm09 bash[21220]: cluster 2026-03-07T06:56:23.096115+0000 mgr.vm07.yrfcuj (mgr.14201) 114 : cluster [DBG] pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:24 vm07 bash[17031]: cluster 2026-03-07T06:56:23.096115+0000 mgr.vm07.yrfcuj (mgr.14201) 114 : cluster [DBG] pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:24.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:24 vm07 bash[17031]: cluster 2026-03-07T06:56:23.096115+0000 mgr.vm07.yrfcuj (mgr.14201) 114 : cluster [DBG] pgmap v70: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:25.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:25 vm09 bash[21220]: audit 2026-03-07T06:56:25.125798+0000 mon.vm07 (mon.0) 631 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:56:25.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:25 vm09 bash[21220]: audit 2026-03-07T06:56:25.125798+0000 mon.vm07 (mon.0) 631 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:56:25.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:25 vm07 bash[17031]: audit 2026-03-07T06:56:25.125798+0000 mon.vm07 (mon.0) 631 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:56:25.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:25 vm07 bash[17031]: audit 2026-03-07T06:56:25.125798+0000 mon.vm07 (mon.0) 631 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:56:26.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:26 vm09 bash[21220]: cluster 2026-03-07T06:56:25.096323+0000 mgr.vm07.yrfcuj (mgr.14201) 115 : cluster [DBG] pgmap v71: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:26.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:26 vm09 bash[21220]: cluster 2026-03-07T06:56:25.096323+0000 mgr.vm07.yrfcuj (mgr.14201) 115 : cluster [DBG] pgmap v71: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:26.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:26 vm07 bash[17031]: cluster 2026-03-07T06:56:25.096323+0000 mgr.vm07.yrfcuj (mgr.14201) 115 : cluster [DBG] pgmap v71: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:26.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:26 vm07 bash[17031]: cluster 2026-03-07T06:56:25.096323+0000 mgr.vm07.yrfcuj (mgr.14201) 115 : cluster [DBG] pgmap v71: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:28.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:28 vm09 bash[21220]: cluster 2026-03-07T06:56:27.096687+0000 mgr.vm07.yrfcuj (mgr.14201) 116 : cluster [DBG] pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:28.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:28 vm09 bash[21220]: cluster 2026-03-07T06:56:27.096687+0000 mgr.vm07.yrfcuj (mgr.14201) 116 : cluster [DBG] pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:28.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:28 vm07 bash[17031]: cluster 2026-03-07T06:56:27.096687+0000 mgr.vm07.yrfcuj (mgr.14201) 116 : cluster [DBG] pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:28.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:28 vm07 bash[17031]: cluster 2026-03-07T06:56:27.096687+0000 mgr.vm07.yrfcuj (mgr.14201) 116 : cluster [DBG] pgmap v72: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:29.223 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:29.226 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:29.227 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:29.229 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:29.231 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:29.232 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:29.233 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:29.233 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:29.788 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:29 vm07 bash[17031]: cluster 2026-03-07T06:56:29.097028+0000 mgr.vm07.yrfcuj (mgr.14201) 117 : cluster [DBG] pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:29.788 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:29 vm07 bash[17031]: cluster 2026-03-07T06:56:29.097028+0000 mgr.vm07.yrfcuj (mgr.14201) 117 : cluster [DBG] pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:29.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:29 vm09 bash[21220]: cluster 2026-03-07T06:56:29.097028+0000 mgr.vm07.yrfcuj (mgr.14201) 117 : cluster [DBG] pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:29.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:29 vm09 bash[21220]: cluster 2026-03-07T06:56:29.097028+0000 mgr.vm07.yrfcuj (mgr.14201) 117 : cluster [DBG] pgmap v73: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:30.299 INFO:teuthology.orchestra.run.vm07.stdout:73014444044 2026-03-07T07:56:30.299 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd last-stat-seq osd.0 2026-03-07T07:56:30.492 INFO:teuthology.orchestra.run.vm07.stdout:90194313227 2026-03-07T07:56:30.492 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd last-stat-seq osd.5 2026-03-07T07:56:30.800 INFO:teuthology.orchestra.run.vm07.stdout:81604378636 2026-03-07T07:56:30.800 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd last-stat-seq osd.3 2026-03-07T07:56:31.007 INFO:teuthology.orchestra.run.vm07.stdout:90194313227 2026-03-07T07:56:31.007 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd last-stat-seq osd.4 2026-03-07T07:56:31.167 INFO:teuthology.orchestra.run.vm07.stdout:90194313227 2026-03-07T07:56:31.167 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd last-stat-seq osd.6 2026-03-07T07:56:31.170 INFO:teuthology.orchestra.run.vm07.stdout:85899345931 2026-03-07T07:56:31.170 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd last-stat-seq osd.2 2026-03-07T07:56:31.198 INFO:teuthology.orchestra.run.vm07.stdout:90194313227 2026-03-07T07:56:31.199 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd last-stat-seq osd.7 2026-03-07T07:56:31.200 INFO:teuthology.orchestra.run.vm07.stdout:68719476748 2026-03-07T07:56:31.200 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph osd last-stat-seq osd.1 2026-03-07T07:56:32.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:32 vm09 bash[21220]: cluster 2026-03-07T06:56:31.097309+0000 mgr.vm07.yrfcuj (mgr.14201) 118 : cluster [DBG] pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:32.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:32 vm09 bash[21220]: cluster 2026-03-07T06:56:31.097309+0000 mgr.vm07.yrfcuj (mgr.14201) 118 : cluster [DBG] pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:32.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:32 vm07 bash[17031]: cluster 2026-03-07T06:56:31.097309+0000 mgr.vm07.yrfcuj (mgr.14201) 118 : cluster [DBG] pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:32.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:32 vm07 bash[17031]: cluster 2026-03-07T06:56:31.097309+0000 mgr.vm07.yrfcuj (mgr.14201) 118 : cluster [DBG] pgmap v74: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:34.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:34 vm09 bash[21220]: cluster 2026-03-07T06:56:33.097634+0000 mgr.vm07.yrfcuj (mgr.14201) 119 : cluster [DBG] pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:34.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:34 vm09 bash[21220]: cluster 2026-03-07T06:56:33.097634+0000 mgr.vm07.yrfcuj (mgr.14201) 119 : cluster [DBG] pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:34.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:34 vm07 bash[17031]: cluster 2026-03-07T06:56:33.097634+0000 mgr.vm07.yrfcuj (mgr.14201) 119 : cluster [DBG] pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:34.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:34 vm07 bash[17031]: cluster 2026-03-07T06:56:33.097634+0000 mgr.vm07.yrfcuj (mgr.14201) 119 : cluster [DBG] pgmap v75: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:35.229 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:35.229 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:35.229 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:35.689 INFO:teuthology.orchestra.run.vm07.stdout:81604378637 2026-03-07T07:56:35.747 INFO:teuthology.orchestra.run.vm07.stdout:73014444045 2026-03-07T07:56:35.807 INFO:tasks.cephadm.ceph_manager.ceph:need seq 81604378636 got 81604378637 for osd.3 2026-03-07T07:56:35.807 DEBUG:teuthology.parallel:result is None 2026-03-07T07:56:35.830 INFO:teuthology.orchestra.run.vm07.stdout:90194313228 2026-03-07T07:56:35.839 INFO:tasks.cephadm.ceph_manager.ceph:need seq 73014444044 got 73014444045 for osd.0 2026-03-07T07:56:35.839 DEBUG:teuthology.parallel:result is None 2026-03-07T07:56:35.914 INFO:tasks.cephadm.ceph_manager.ceph:need seq 90194313227 got 90194313228 for osd.5 2026-03-07T07:56:35.914 DEBUG:teuthology.parallel:result is None 2026-03-07T07:56:36.392 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:36.392 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:36.392 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:36.392 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:36.394 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:36.458 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:36 vm07 bash[17031]: cluster 2026-03-07T06:56:35.097899+0000 mgr.vm07.yrfcuj (mgr.14201) 120 : cluster [DBG] pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:36.458 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:36 vm07 bash[17031]: cluster 2026-03-07T06:56:35.097899+0000 mgr.vm07.yrfcuj (mgr.14201) 120 : cluster [DBG] pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:36.458 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:36 vm07 bash[17031]: audit 2026-03-07T06:56:35.683404+0000 mon.vm07 (mon.0) 632 : audit [DBG] from='client.? 192.168.123.107:0/3876146308' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-07T07:56:36.458 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:36 vm07 bash[17031]: audit 2026-03-07T06:56:35.683404+0000 mon.vm07 (mon.0) 632 : audit [DBG] from='client.? 192.168.123.107:0/3876146308' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-07T07:56:36.458 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:36 vm07 bash[17031]: audit 2026-03-07T06:56:35.738811+0000 mon.vm07 (mon.0) 633 : audit [DBG] from='client.? 192.168.123.107:0/4241098937' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-07T07:56:36.458 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:36 vm07 bash[17031]: audit 2026-03-07T06:56:35.738811+0000 mon.vm07 (mon.0) 633 : audit [DBG] from='client.? 192.168.123.107:0/4241098937' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-07T07:56:36.458 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:36 vm07 bash[17031]: audit 2026-03-07T06:56:35.823243+0000 mon.vm07 (mon.0) 634 : audit [DBG] from='client.? 192.168.123.107:0/3475393162' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-07T07:56:36.458 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:36 vm07 bash[17031]: audit 2026-03-07T06:56:35.823243+0000 mon.vm07 (mon.0) 634 : audit [DBG] from='client.? 192.168.123.107:0/3475393162' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-07T07:56:36.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:36 vm09 bash[21220]: cluster 2026-03-07T06:56:35.097899+0000 mgr.vm07.yrfcuj (mgr.14201) 120 : cluster [DBG] pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:36.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:36 vm09 bash[21220]: cluster 2026-03-07T06:56:35.097899+0000 mgr.vm07.yrfcuj (mgr.14201) 120 : cluster [DBG] pgmap v76: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:36.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:36 vm09 bash[21220]: audit 2026-03-07T06:56:35.683404+0000 mon.vm07 (mon.0) 632 : audit [DBG] from='client.? 192.168.123.107:0/3876146308' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-07T07:56:36.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:36 vm09 bash[21220]: audit 2026-03-07T06:56:35.683404+0000 mon.vm07 (mon.0) 632 : audit [DBG] from='client.? 192.168.123.107:0/3876146308' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-07T07:56:36.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:36 vm09 bash[21220]: audit 2026-03-07T06:56:35.738811+0000 mon.vm07 (mon.0) 633 : audit [DBG] from='client.? 192.168.123.107:0/4241098937' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-07T07:56:36.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:36 vm09 bash[21220]: audit 2026-03-07T06:56:35.738811+0000 mon.vm07 (mon.0) 633 : audit [DBG] from='client.? 192.168.123.107:0/4241098937' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-07T07:56:36.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:36 vm09 bash[21220]: audit 2026-03-07T06:56:35.823243+0000 mon.vm07 (mon.0) 634 : audit [DBG] from='client.? 192.168.123.107:0/3475393162' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-07T07:56:36.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:36 vm09 bash[21220]: audit 2026-03-07T06:56:35.823243+0000 mon.vm07 (mon.0) 634 : audit [DBG] from='client.? 192.168.123.107:0/3475393162' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-07T07:56:37.140 INFO:teuthology.orchestra.run.vm07.stdout:85899345932 2026-03-07T07:56:37.188 INFO:teuthology.orchestra.run.vm07.stdout:90194313228 2026-03-07T07:56:37.255 INFO:tasks.cephadm.ceph_manager.ceph:need seq 85899345931 got 85899345932 for osd.2 2026-03-07T07:56:37.255 DEBUG:teuthology.parallel:result is None 2026-03-07T07:56:37.284 INFO:teuthology.orchestra.run.vm07.stdout:90194313229 2026-03-07T07:56:37.293 INFO:tasks.cephadm.ceph_manager.ceph:need seq 90194313227 got 90194313228 for osd.4 2026-03-07T07:56:37.293 DEBUG:teuthology.parallel:result is None 2026-03-07T07:56:37.300 INFO:teuthology.orchestra.run.vm07.stdout:90194313229 2026-03-07T07:56:37.319 INFO:teuthology.orchestra.run.vm07.stdout:68719476749 2026-03-07T07:56:37.374 INFO:tasks.cephadm.ceph_manager.ceph:need seq 90194313227 got 90194313229 for osd.7 2026-03-07T07:56:37.374 DEBUG:teuthology.parallel:result is None 2026-03-07T07:56:37.431 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:37 vm07 bash[17031]: audit 2026-03-07T06:56:37.129212+0000 mon.vm07 (mon.0) 635 : audit [DBG] from='client.? 192.168.123.107:0/2721392323' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-07T07:56:37.431 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:37 vm07 bash[17031]: audit 2026-03-07T06:56:37.129212+0000 mon.vm07 (mon.0) 635 : audit [DBG] from='client.? 192.168.123.107:0/2721392323' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-07T07:56:37.431 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:37 vm07 bash[17031]: audit 2026-03-07T06:56:37.174570+0000 mon.vm07 (mon.0) 636 : audit [DBG] from='client.? 192.168.123.107:0/1107812821' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-07T07:56:37.431 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:37 vm07 bash[17031]: audit 2026-03-07T06:56:37.174570+0000 mon.vm07 (mon.0) 636 : audit [DBG] from='client.? 192.168.123.107:0/1107812821' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-07T07:56:37.431 INFO:tasks.cephadm.ceph_manager.ceph:need seq 90194313227 got 90194313229 for osd.6 2026-03-07T07:56:37.431 DEBUG:teuthology.parallel:result is None 2026-03-07T07:56:37.443 INFO:tasks.cephadm.ceph_manager.ceph:need seq 68719476748 got 68719476749 for osd.1 2026-03-07T07:56:37.443 DEBUG:teuthology.parallel:result is None 2026-03-07T07:56:37.443 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-07T07:56:37.444 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph pg dump --format=json 2026-03-07T07:56:37.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:37 vm09 bash[21220]: audit 2026-03-07T06:56:37.129212+0000 mon.vm07 (mon.0) 635 : audit [DBG] from='client.? 192.168.123.107:0/2721392323' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-07T07:56:37.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:37 vm09 bash[21220]: audit 2026-03-07T06:56:37.129212+0000 mon.vm07 (mon.0) 635 : audit [DBG] from='client.? 192.168.123.107:0/2721392323' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-07T07:56:37.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:37 vm09 bash[21220]: audit 2026-03-07T06:56:37.174570+0000 mon.vm07 (mon.0) 636 : audit [DBG] from='client.? 192.168.123.107:0/1107812821' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-07T07:56:37.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:37 vm09 bash[21220]: audit 2026-03-07T06:56:37.174570+0000 mon.vm07 (mon.0) 636 : audit [DBG] from='client.? 192.168.123.107:0/1107812821' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-07T07:56:38.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:38 vm09 bash[21220]: cluster 2026-03-07T06:56:37.098177+0000 mgr.vm07.yrfcuj (mgr.14201) 121 : cluster [DBG] pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:38.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:38 vm09 bash[21220]: cluster 2026-03-07T06:56:37.098177+0000 mgr.vm07.yrfcuj (mgr.14201) 121 : cluster [DBG] pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:38.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:38 vm09 bash[21220]: audit 2026-03-07T06:56:37.267553+0000 mon.vm09 (mon.1) 20 : audit [DBG] from='client.? 192.168.123.107:0/1441474230' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-07T07:56:38.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:38 vm09 bash[21220]: audit 2026-03-07T06:56:37.267553+0000 mon.vm09 (mon.1) 20 : audit [DBG] from='client.? 192.168.123.107:0/1441474230' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-07T07:56:38.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:38 vm09 bash[21220]: audit 2026-03-07T06:56:37.289664+0000 mon.vm07 (mon.0) 637 : audit [DBG] from='client.? 192.168.123.107:0/2768534998' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-07T07:56:38.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:38 vm09 bash[21220]: audit 2026-03-07T06:56:37.289664+0000 mon.vm07 (mon.0) 637 : audit [DBG] from='client.? 192.168.123.107:0/2768534998' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-07T07:56:38.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:38 vm09 bash[21220]: audit 2026-03-07T06:56:37.313482+0000 mon.vm07 (mon.0) 638 : audit [DBG] from='client.? 192.168.123.107:0/4125367524' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-07T07:56:38.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:38 vm09 bash[21220]: audit 2026-03-07T06:56:37.313482+0000 mon.vm07 (mon.0) 638 : audit [DBG] from='client.? 192.168.123.107:0/4125367524' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-07T07:56:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:38 vm07 bash[17031]: cluster 2026-03-07T06:56:37.098177+0000 mgr.vm07.yrfcuj (mgr.14201) 121 : cluster [DBG] pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:38 vm07 bash[17031]: cluster 2026-03-07T06:56:37.098177+0000 mgr.vm07.yrfcuj (mgr.14201) 121 : cluster [DBG] pgmap v77: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:38 vm07 bash[17031]: audit 2026-03-07T06:56:37.267553+0000 mon.vm09 (mon.1) 20 : audit [DBG] from='client.? 192.168.123.107:0/1441474230' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-07T07:56:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:38 vm07 bash[17031]: audit 2026-03-07T06:56:37.267553+0000 mon.vm09 (mon.1) 20 : audit [DBG] from='client.? 192.168.123.107:0/1441474230' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-07T07:56:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:38 vm07 bash[17031]: audit 2026-03-07T06:56:37.289664+0000 mon.vm07 (mon.0) 637 : audit [DBG] from='client.? 192.168.123.107:0/2768534998' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-07T07:56:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:38 vm07 bash[17031]: audit 2026-03-07T06:56:37.289664+0000 mon.vm07 (mon.0) 637 : audit [DBG] from='client.? 192.168.123.107:0/2768534998' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-07T07:56:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:38 vm07 bash[17031]: audit 2026-03-07T06:56:37.313482+0000 mon.vm07 (mon.0) 638 : audit [DBG] from='client.? 192.168.123.107:0/4125367524' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-07T07:56:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:38 vm07 bash[17031]: audit 2026-03-07T06:56:37.313482+0000 mon.vm07 (mon.0) 638 : audit [DBG] from='client.? 192.168.123.107:0/4125367524' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-07T07:56:40.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:40 vm09 bash[21220]: cluster 2026-03-07T06:56:39.098461+0000 mgr.vm07.yrfcuj (mgr.14201) 122 : cluster [DBG] pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:40.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:40 vm09 bash[21220]: cluster 2026-03-07T06:56:39.098461+0000 mgr.vm07.yrfcuj (mgr.14201) 122 : cluster [DBG] pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:40.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:40 vm09 bash[21220]: audit 2026-03-07T06:56:40.126352+0000 mon.vm07 (mon.0) 639 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:56:40.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:40 vm09 bash[21220]: audit 2026-03-07T06:56:40.126352+0000 mon.vm07 (mon.0) 639 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:56:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:40 vm07 bash[17031]: cluster 2026-03-07T06:56:39.098461+0000 mgr.vm07.yrfcuj (mgr.14201) 122 : cluster [DBG] pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:40 vm07 bash[17031]: cluster 2026-03-07T06:56:39.098461+0000 mgr.vm07.yrfcuj (mgr.14201) 122 : cluster [DBG] pgmap v78: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:40 vm07 bash[17031]: audit 2026-03-07T06:56:40.126352+0000 mon.vm07 (mon.0) 639 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:56:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:40 vm07 bash[17031]: audit 2026-03-07T06:56:40.126352+0000 mon.vm07 (mon.0) 639 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:56:42.263 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:42.605 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:56:42.605 INFO:teuthology.orchestra.run.vm07.stderr:dumped all 2026-03-07T07:56:42.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:42 vm09 bash[21220]: cluster 2026-03-07T06:56:41.098734+0000 mgr.vm07.yrfcuj (mgr.14201) 123 : cluster [DBG] pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:42.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:42 vm09 bash[21220]: cluster 2026-03-07T06:56:41.098734+0000 mgr.vm07.yrfcuj (mgr.14201) 123 : cluster [DBG] pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:42.616 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:42 vm07 bash[17031]: cluster 2026-03-07T06:56:41.098734+0000 mgr.vm07.yrfcuj (mgr.14201) 123 : cluster [DBG] pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:42.616 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:42 vm07 bash[17031]: cluster 2026-03-07T06:56:41.098734+0000 mgr.vm07.yrfcuj (mgr.14201) 123 : cluster [DBG] pgmap v79: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:42.668 INFO:teuthology.orchestra.run.vm07.stdout:{"pg_ready":true,"pg_map":{"version":79,"stamp":"2026-03-07T06:56:41.098604+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":3,"kb":167739392,"kb_used":218108,"kb_used_data":2988,"kb_used_omap":12,"kb_used_meta":214515,"kb_avail":167521284,"statfs":{"total":171765137408,"available":171541794816,"internally_reserved":0,"allocated":3059712,"data_stored":1986552,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":12712,"internal_metadata":219663960},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"12.001717"},"pg_stats":[{"pgid":"1.0","version":"22'32","reported_seq":57,"reported_epoch":23,"state":"active+clean","last_fresh":"2026-03-07T06:55:45.081241+0000","last_change":"2026-03-07T06:55:43.270311+0000","last_active":"2026-03-07T06:55:45.081241+0000","last_peered":"2026-03-07T06:55:45.081241+0000","last_clean":"2026-03-07T06:55:45.081241+0000","last_became_active":"2026-03-07T06:55:43.269592+0000","last_became_peered":"2026-03-07T06:55:43.269592+0000","last_unstale":"2026-03-07T06:55:45.081241+0000","last_undegraded":"2026-03-07T06:55:45.081241+0000","last_fullsized":"2026-03-07T06:55:45.081241+0000","mapping_epoch":21,"log_start":"0'0","ondisk_log_start":"0'0","created":21,"last_epoch_clean":22,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-07T06:55:41.609166+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-07T06:55:41.609166+0000","last_clean_scrub_stamp":"2026-03-07T06:55:41.609166+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-08T17:05:04.090991+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,5,3],"acting":[7,5,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1388544,"data_stored":1377840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":7,"up_from":21,"seq":90194313229,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27624,"kb_used_data":656,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939800,"statfs":{"total":21470642176,"available":21442355200,"internally_reserved":0,"allocated":671744,"data_stored":535369,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1585,"internal_metadata":27457999},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":21,"seq":90194313229,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27044,"kb_used_data":204,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940380,"statfs":{"total":21470642176,"available":21442949120,"internally_reserved":0,"allocated":208896,"data_stored":76089,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":21,"seq":90194313229,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27628,"kb_used_data":656,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939796,"statfs":{"total":21470642176,"available":21442351104,"internally_reserved":0,"allocated":671744,"data_stored":535369,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":21,"seq":90194313229,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27048,"kb_used_data":204,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940376,"statfs":{"total":21470642176,"available":21442945024,"internally_reserved":0,"allocated":208896,"data_stored":76089,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":20,"seq":85899345933,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27044,"kb_used_data":204,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940380,"statfs":{"total":21470642176,"available":21442949120,"internally_reserved":0,"allocated":208896,"data_stored":76089,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":19,"seq":81604378638,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27624,"kb_used_data":656,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939800,"statfs":{"total":21470642176,"available":21442355200,"internally_reserved":0,"allocated":671744,"data_stored":535369,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":17,"seq":73014444046,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27044,"kb_used_data":204,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940380,"statfs":{"total":21470642176,"available":21442949120,"internally_reserved":0,"allocated":208896,"data_stored":76089,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":16,"seq":68719476750,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27052,"kb_used_data":204,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940372,"statfs":{"total":21470642176,"available":21442940928,"internally_reserved":0,"allocated":208896,"data_stored":76089,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-07T07:56:42.668 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph pg dump --format=json 2026-03-07T07:56:44.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:44 vm09 bash[21220]: audit 2026-03-07T06:56:42.599783+0000 mgr.vm07.yrfcuj (mgr.14201) 124 : audit [DBG] from='client.14448 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:56:44.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:44 vm09 bash[21220]: audit 2026-03-07T06:56:42.599783+0000 mgr.vm07.yrfcuj (mgr.14201) 124 : audit [DBG] from='client.14448 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:56:44.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:44 vm09 bash[21220]: cluster 2026-03-07T06:56:43.099091+0000 mgr.vm07.yrfcuj (mgr.14201) 125 : cluster [DBG] pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:44.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:44 vm09 bash[21220]: cluster 2026-03-07T06:56:43.099091+0000 mgr.vm07.yrfcuj (mgr.14201) 125 : cluster [DBG] pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:44 vm07 bash[17031]: audit 2026-03-07T06:56:42.599783+0000 mgr.vm07.yrfcuj (mgr.14201) 124 : audit [DBG] from='client.14448 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:56:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:44 vm07 bash[17031]: audit 2026-03-07T06:56:42.599783+0000 mgr.vm07.yrfcuj (mgr.14201) 124 : audit [DBG] from='client.14448 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:56:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:44 vm07 bash[17031]: cluster 2026-03-07T06:56:43.099091+0000 mgr.vm07.yrfcuj (mgr.14201) 125 : cluster [DBG] pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:44 vm07 bash[17031]: cluster 2026-03-07T06:56:43.099091+0000 mgr.vm07.yrfcuj (mgr.14201) 125 : cluster [DBG] pgmap v80: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:46.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:46 vm07 bash[17031]: cluster 2026-03-07T06:56:45.099328+0000 mgr.vm07.yrfcuj (mgr.14201) 126 : cluster [DBG] pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:46.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:46 vm07 bash[17031]: cluster 2026-03-07T06:56:45.099328+0000 mgr.vm07.yrfcuj (mgr.14201) 126 : cluster [DBG] pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:46.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:46 vm09 bash[21220]: cluster 2026-03-07T06:56:45.099328+0000 mgr.vm07.yrfcuj (mgr.14201) 126 : cluster [DBG] pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:46.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:46 vm09 bash[21220]: cluster 2026-03-07T06:56:45.099328+0000 mgr.vm07.yrfcuj (mgr.14201) 126 : cluster [DBG] pgmap v81: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:47.446 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:47.812 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:56:47.812 INFO:teuthology.orchestra.run.vm07.stderr:dumped all 2026-03-07T07:56:47.874 INFO:teuthology.orchestra.run.vm07.stdout:{"pg_ready":true,"pg_map":{"version":82,"stamp":"2026-03-07T06:56:47.099467+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":3,"kb":167739392,"kb_used":218108,"kb_used_data":2988,"kb_used_omap":12,"kb_used_meta":214515,"kb_avail":167521284,"statfs":{"total":171765137408,"available":171541794816,"internally_reserved":0,"allocated":3059712,"data_stored":1986552,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":12712,"internal_metadata":219663960},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"12.001694"},"pg_stats":[{"pgid":"1.0","version":"22'32","reported_seq":57,"reported_epoch":23,"state":"active+clean","last_fresh":"2026-03-07T06:55:45.081241+0000","last_change":"2026-03-07T06:55:43.270311+0000","last_active":"2026-03-07T06:55:45.081241+0000","last_peered":"2026-03-07T06:55:45.081241+0000","last_clean":"2026-03-07T06:55:45.081241+0000","last_became_active":"2026-03-07T06:55:43.269592+0000","last_became_peered":"2026-03-07T06:55:43.269592+0000","last_unstale":"2026-03-07T06:55:45.081241+0000","last_undegraded":"2026-03-07T06:55:45.081241+0000","last_fullsized":"2026-03-07T06:55:45.081241+0000","mapping_epoch":21,"log_start":"0'0","ondisk_log_start":"0'0","created":21,"last_epoch_clean":22,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-07T06:55:41.609166+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-07T06:55:41.609166+0000","last_clean_scrub_stamp":"2026-03-07T06:55:41.609166+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-08T17:05:04.090991+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,5,3],"acting":[7,5,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1388544,"data_stored":1377840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":7,"up_from":21,"seq":90194313231,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27624,"kb_used_data":656,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939800,"statfs":{"total":21470642176,"available":21442355200,"internally_reserved":0,"allocated":671744,"data_stored":535369,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1585,"internal_metadata":27457999},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":21,"seq":90194313231,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27044,"kb_used_data":204,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940380,"statfs":{"total":21470642176,"available":21442949120,"internally_reserved":0,"allocated":208896,"data_stored":76089,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":21,"seq":90194313230,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27628,"kb_used_data":656,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939796,"statfs":{"total":21470642176,"available":21442351104,"internally_reserved":0,"allocated":671744,"data_stored":535369,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":21,"seq":90194313230,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27048,"kb_used_data":204,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940376,"statfs":{"total":21470642176,"available":21442945024,"internally_reserved":0,"allocated":208896,"data_stored":76089,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":20,"seq":85899345935,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27044,"kb_used_data":204,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940380,"statfs":{"total":21470642176,"available":21442949120,"internally_reserved":0,"allocated":208896,"data_stored":76089,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":19,"seq":81604378639,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27624,"kb_used_data":656,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939800,"statfs":{"total":21470642176,"available":21442355200,"internally_reserved":0,"allocated":671744,"data_stored":535369,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":17,"seq":73014444047,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27044,"kb_used_data":204,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940380,"statfs":{"total":21470642176,"available":21442949120,"internally_reserved":0,"allocated":208896,"data_stored":76089,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":16,"seq":68719476751,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27052,"kb_used_data":204,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940372,"statfs":{"total":21470642176,"available":21442940928,"internally_reserved":0,"allocated":208896,"data_stored":76089,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-07T07:56:47.874 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-07T07:56:47.874 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-07T07:56:47.874 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-07T07:56:47.874 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph health --format=json 2026-03-07T07:56:48.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:48 vm09 bash[21220]: cluster 2026-03-07T06:56:47.099587+0000 mgr.vm07.yrfcuj (mgr.14201) 127 : cluster [DBG] pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:48.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:48 vm09 bash[21220]: cluster 2026-03-07T06:56:47.099587+0000 mgr.vm07.yrfcuj (mgr.14201) 127 : cluster [DBG] pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:48 vm07 bash[17031]: cluster 2026-03-07T06:56:47.099587+0000 mgr.vm07.yrfcuj (mgr.14201) 127 : cluster [DBG] pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:48.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:48 vm07 bash[17031]: cluster 2026-03-07T06:56:47.099587+0000 mgr.vm07.yrfcuj (mgr.14201) 127 : cluster [DBG] pgmap v82: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:49.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:49 vm09 bash[21220]: audit 2026-03-07T06:56:47.806273+0000 mgr.vm07.yrfcuj (mgr.14201) 128 : audit [DBG] from='client.14452 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:56:49.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:49 vm09 bash[21220]: audit 2026-03-07T06:56:47.806273+0000 mgr.vm07.yrfcuj (mgr.14201) 128 : audit [DBG] from='client.14452 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:56:49.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:49 vm07 bash[17031]: audit 2026-03-07T06:56:47.806273+0000 mgr.vm07.yrfcuj (mgr.14201) 128 : audit [DBG] from='client.14452 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:56:49.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:49 vm07 bash[17031]: audit 2026-03-07T06:56:47.806273+0000 mgr.vm07.yrfcuj (mgr.14201) 128 : audit [DBG] from='client.14452 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:56:50.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:50 vm09 bash[21220]: cluster 2026-03-07T06:56:49.099898+0000 mgr.vm07.yrfcuj (mgr.14201) 129 : cluster [DBG] pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:50.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:50 vm09 bash[21220]: cluster 2026-03-07T06:56:49.099898+0000 mgr.vm07.yrfcuj (mgr.14201) 129 : cluster [DBG] pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:50 vm07 bash[17031]: cluster 2026-03-07T06:56:49.099898+0000 mgr.vm07.yrfcuj (mgr.14201) 129 : cluster [DBG] pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:50 vm07 bash[17031]: cluster 2026-03-07T06:56:49.099898+0000 mgr.vm07.yrfcuj (mgr.14201) 129 : cluster [DBG] pgmap v83: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:52.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:52 vm09 bash[21220]: cluster 2026-03-07T06:56:51.100148+0000 mgr.vm07.yrfcuj (mgr.14201) 130 : cluster [DBG] pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:52.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:52 vm09 bash[21220]: cluster 2026-03-07T06:56:51.100148+0000 mgr.vm07.yrfcuj (mgr.14201) 130 : cluster [DBG] pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:52.656 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:52.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:52 vm07 bash[17031]: cluster 2026-03-07T06:56:51.100148+0000 mgr.vm07.yrfcuj (mgr.14201) 130 : cluster [DBG] pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:52.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:52 vm07 bash[17031]: cluster 2026-03-07T06:56:51.100148+0000 mgr.vm07.yrfcuj (mgr.14201) 130 : cluster [DBG] pgmap v84: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:53.035 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:56:53.035 INFO:teuthology.orchestra.run.vm07.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-07T07:56:53.095 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-07T07:56:53.095 INFO:tasks.cephadm:Setup complete, yielding 2026-03-07T07:56:53.095 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-07T07:56:53.097 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm07.local 2026-03-07T07:56:53.097 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- bash -c 'ceph orch status' 2026-03-07T07:56:53.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:53 vm07 bash[17031]: audit 2026-03-07T06:56:53.030169+0000 mon.vm07 (mon.0) 640 : audit [DBG] from='client.? 192.168.123.107:0/1667226531' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-07T07:56:53.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:53 vm07 bash[17031]: audit 2026-03-07T06:56:53.030169+0000 mon.vm07 (mon.0) 640 : audit [DBG] from='client.? 192.168.123.107:0/1667226531' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-07T07:56:53.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:53 vm09 bash[21220]: audit 2026-03-07T06:56:53.030169+0000 mon.vm07 (mon.0) 640 : audit [DBG] from='client.? 192.168.123.107:0/1667226531' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-07T07:56:53.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:53 vm09 bash[21220]: audit 2026-03-07T06:56:53.030169+0000 mon.vm07 (mon.0) 640 : audit [DBG] from='client.? 192.168.123.107:0/1667226531' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-07T07:56:54.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:54 vm09 bash[21220]: cluster 2026-03-07T06:56:53.100420+0000 mgr.vm07.yrfcuj (mgr.14201) 131 : cluster [DBG] pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:54.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:54 vm09 bash[21220]: cluster 2026-03-07T06:56:53.100420+0000 mgr.vm07.yrfcuj (mgr.14201) 131 : cluster [DBG] pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:54.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:54 vm07 bash[17031]: cluster 2026-03-07T06:56:53.100420+0000 mgr.vm07.yrfcuj (mgr.14201) 131 : cluster [DBG] pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:54.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:54 vm07 bash[17031]: cluster 2026-03-07T06:56:53.100420+0000 mgr.vm07.yrfcuj (mgr.14201) 131 : cluster [DBG] pgmap v85: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:55.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:55 vm09 bash[21220]: audit 2026-03-07T06:56:54.923500+0000 mon.vm07 (mon.0) 641 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:56:55.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:55 vm09 bash[21220]: audit 2026-03-07T06:56:54.923500+0000 mon.vm07 (mon.0) 641 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:56:55.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:55 vm09 bash[21220]: audit 2026-03-07T06:56:55.126378+0000 mon.vm07 (mon.0) 642 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:56:55.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:55 vm09 bash[21220]: audit 2026-03-07T06:56:55.126378+0000 mon.vm07 (mon.0) 642 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:56:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:55 vm07 bash[17031]: audit 2026-03-07T06:56:54.923500+0000 mon.vm07 (mon.0) 641 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:56:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:55 vm07 bash[17031]: audit 2026-03-07T06:56:54.923500+0000 mon.vm07 (mon.0) 641 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:56:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:55 vm07 bash[17031]: audit 2026-03-07T06:56:55.126378+0000 mon.vm07 (mon.0) 642 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:56:55.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:55 vm07 bash[17031]: audit 2026-03-07T06:56:55.126378+0000 mon.vm07 (mon.0) 642 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:56:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:56 vm07 bash[17031]: cluster 2026-03-07T06:56:55.100701+0000 mgr.vm07.yrfcuj (mgr.14201) 132 : cluster [DBG] pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:56 vm07 bash[17031]: cluster 2026-03-07T06:56:55.100701+0000 mgr.vm07.yrfcuj (mgr.14201) 132 : cluster [DBG] pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:56.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:56 vm09 bash[21220]: cluster 2026-03-07T06:56:55.100701+0000 mgr.vm07.yrfcuj (mgr.14201) 132 : cluster [DBG] pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:56.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:56 vm09 bash[21220]: cluster 2026-03-07T06:56:55.100701+0000 mgr.vm07.yrfcuj (mgr.14201) 132 : cluster [DBG] pgmap v86: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:57.897 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:56:58.263 INFO:teuthology.orchestra.run.vm07.stdout:Backend: cephadm 2026-03-07T07:56:58.263 INFO:teuthology.orchestra.run.vm07.stdout:Available: Yes 2026-03-07T07:56:58.263 INFO:teuthology.orchestra.run.vm07.stdout:Paused: No 2026-03-07T07:56:58.329 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- bash -c 'ceph orch ps' 2026-03-07T07:56:58.345 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:58 vm09 bash[21220]: cluster 2026-03-07T06:56:57.100996+0000 mgr.vm07.yrfcuj (mgr.14201) 133 : cluster [DBG] pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:58.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:58 vm09 bash[21220]: cluster 2026-03-07T06:56:57.100996+0000 mgr.vm07.yrfcuj (mgr.14201) 133 : cluster [DBG] pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:58.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:58 vm07 bash[17031]: cluster 2026-03-07T06:56:57.100996+0000 mgr.vm07.yrfcuj (mgr.14201) 133 : cluster [DBG] pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:58.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:58 vm07 bash[17031]: cluster 2026-03-07T06:56:57.100996+0000 mgr.vm07.yrfcuj (mgr.14201) 133 : cluster [DBG] pgmap v87: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:56:59.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:59 vm09 bash[21220]: audit 2026-03-07T06:56:58.257694+0000 mgr.vm07.yrfcuj (mgr.14201) 134 : audit [DBG] from='client.14460 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:56:59.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:56:59 vm09 bash[21220]: audit 2026-03-07T06:56:58.257694+0000 mgr.vm07.yrfcuj (mgr.14201) 134 : audit [DBG] from='client.14460 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:56:59.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:59 vm07 bash[17031]: audit 2026-03-07T06:56:58.257694+0000 mgr.vm07.yrfcuj (mgr.14201) 134 : audit [DBG] from='client.14460 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:56:59.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:56:59 vm07 bash[17031]: audit 2026-03-07T06:56:58.257694+0000 mgr.vm07.yrfcuj (mgr.14201) 134 : audit [DBG] from='client.14460 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:00.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:00 vm07 bash[17031]: cluster 2026-03-07T06:56:59.101229+0000 mgr.vm07.yrfcuj (mgr.14201) 135 : cluster [DBG] pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:00.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:00 vm07 bash[17031]: cluster 2026-03-07T06:56:59.101229+0000 mgr.vm07.yrfcuj (mgr.14201) 135 : cluster [DBG] pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:00.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:00 vm07 bash[17031]: audit 2026-03-07T06:56:59.508762+0000 mon.vm07 (mon.0) 643 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:00.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:00 vm07 bash[17031]: audit 2026-03-07T06:56:59.508762+0000 mon.vm07 (mon.0) 643 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:00.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:00 vm07 bash[17031]: audit 2026-03-07T06:56:59.513456+0000 mon.vm07 (mon.0) 644 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:00.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:00 vm07 bash[17031]: audit 2026-03-07T06:56:59.513456+0000 mon.vm07 (mon.0) 644 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:00.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:00 vm07 bash[17031]: audit 2026-03-07T06:57:00.047806+0000 mon.vm07 (mon.0) 645 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:00.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:00 vm07 bash[17031]: audit 2026-03-07T06:57:00.047806+0000 mon.vm07 (mon.0) 645 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:00.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:00 vm07 bash[17031]: audit 2026-03-07T06:57:00.052854+0000 mon.vm07 (mon.0) 646 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:00.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:00 vm07 bash[17031]: audit 2026-03-07T06:57:00.052854+0000 mon.vm07 (mon.0) 646 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:00.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:00 vm09 bash[21220]: cluster 2026-03-07T06:56:59.101229+0000 mgr.vm07.yrfcuj (mgr.14201) 135 : cluster [DBG] pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:00.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:00 vm09 bash[21220]: cluster 2026-03-07T06:56:59.101229+0000 mgr.vm07.yrfcuj (mgr.14201) 135 : cluster [DBG] pgmap v88: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:00.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:00 vm09 bash[21220]: audit 2026-03-07T06:56:59.508762+0000 mon.vm07 (mon.0) 643 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:00.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:00 vm09 bash[21220]: audit 2026-03-07T06:56:59.508762+0000 mon.vm07 (mon.0) 643 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:00.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:00 vm09 bash[21220]: audit 2026-03-07T06:56:59.513456+0000 mon.vm07 (mon.0) 644 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:00 vm09 bash[21220]: audit 2026-03-07T06:56:59.513456+0000 mon.vm07 (mon.0) 644 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:00 vm09 bash[21220]: audit 2026-03-07T06:57:00.047806+0000 mon.vm07 (mon.0) 645 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:00 vm09 bash[21220]: audit 2026-03-07T06:57:00.047806+0000 mon.vm07 (mon.0) 645 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:00 vm09 bash[21220]: audit 2026-03-07T06:57:00.052854+0000 mon.vm07 (mon.0) 646 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:00 vm09 bash[21220]: audit 2026-03-07T06:57:00.052854+0000 mon.vm07 (mon.0) 646 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:01.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:01 vm07 bash[17031]: audit 2026-03-07T06:57:00.387866+0000 mon.vm07 (mon.0) 647 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:01.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:01 vm07 bash[17031]: audit 2026-03-07T06:57:00.387866+0000 mon.vm07 (mon.0) 647 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:01.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:01 vm07 bash[17031]: audit 2026-03-07T06:57:00.388702+0000 mon.vm07 (mon.0) 648 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:57:01.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:01 vm07 bash[17031]: audit 2026-03-07T06:57:00.388702+0000 mon.vm07 (mon.0) 648 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:57:01.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:01 vm07 bash[17031]: audit 2026-03-07T06:57:00.396468+0000 mon.vm07 (mon.0) 649 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:01.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:01 vm07 bash[17031]: audit 2026-03-07T06:57:00.396468+0000 mon.vm07 (mon.0) 649 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:01.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:01 vm07 bash[17031]: audit 2026-03-07T06:57:00.398771+0000 mon.vm07 (mon.0) 650 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:57:01.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:01 vm07 bash[17031]: audit 2026-03-07T06:57:00.398771+0000 mon.vm07 (mon.0) 650 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:57:01.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:01 vm09 bash[21220]: audit 2026-03-07T06:57:00.387866+0000 mon.vm07 (mon.0) 647 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:01.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:01 vm09 bash[21220]: audit 2026-03-07T06:57:00.387866+0000 mon.vm07 (mon.0) 647 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:01.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:01 vm09 bash[21220]: audit 2026-03-07T06:57:00.388702+0000 mon.vm07 (mon.0) 648 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:57:01.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:01 vm09 bash[21220]: audit 2026-03-07T06:57:00.388702+0000 mon.vm07 (mon.0) 648 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:57:01.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:01 vm09 bash[21220]: audit 2026-03-07T06:57:00.396468+0000 mon.vm07 (mon.0) 649 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:01.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:01 vm09 bash[21220]: audit 2026-03-07T06:57:00.396468+0000 mon.vm07 (mon.0) 649 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:01.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:01 vm09 bash[21220]: audit 2026-03-07T06:57:00.398771+0000 mon.vm07 (mon.0) 650 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:57:01.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:01 vm09 bash[21220]: audit 2026-03-07T06:57:00.398771+0000 mon.vm07 (mon.0) 650 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:57:02.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:02 vm07 bash[17031]: cluster 2026-03-07T06:57:01.101498+0000 mgr.vm07.yrfcuj (mgr.14201) 136 : cluster [DBG] pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:02.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:02 vm07 bash[17031]: cluster 2026-03-07T06:57:01.101498+0000 mgr.vm07.yrfcuj (mgr.14201) 136 : cluster [DBG] pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:02.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:02 vm09 bash[21220]: cluster 2026-03-07T06:57:01.101498+0000 mgr.vm07.yrfcuj (mgr.14201) 136 : cluster [DBG] pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:02.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:02 vm09 bash[21220]: cluster 2026-03-07T06:57:01.101498+0000 mgr.vm07.yrfcuj (mgr.14201) 136 : cluster [DBG] pgmap v89: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:02.945 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:alertmanager.vm07 vm07 *:9093,9094 running (2m) 3s ago 3m 14.4M - 0.25.0 c8568f914cd2 ced8afb39df2 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:ceph-exporter.vm07 vm07 running (3m) 3s ago 3m 8664k - 19.2.3-39-g340d3c24fc6 8bccc98d839a 65356e296978 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:ceph-exporter.vm09 vm09 running (2m) 3s ago 2m 8047k - 19.2.3-39-g340d3c24fc6 8bccc98d839a ef35692c433d 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:crash.vm07 vm07 running (3m) 3s ago 3m 10.7M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c38c18a5b965 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:crash.vm09 vm09 running (2m) 3s ago 2m 10.7M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 16d01acf9cbb 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:grafana.vm07 vm07 *:3000 running (2m) 3s ago 2m 64.4M - 10.4.0 c8b91775d855 fbed394a229c 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:mgr.vm07.yrfcuj vm07 *:9283,8765,8443 running (4m) 3s ago 4m 526M - 19.2.3-39-g340d3c24fc6 8bccc98d839a afb3597f2ccd 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:mgr.vm09.eqznpw vm09 *:8443,9283,8765 running (2m) 3s ago 2m 471M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 046b6d4953d0 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:mon.vm07 vm07 running (4m) 3s ago 4m 49.0M 2048M 19.2.3-39-g340d3c24fc6 8bccc98d839a df7f9a962253 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:mon.vm09 vm09 running (2m) 3s ago 2m 40.4M 2048M 19.2.3-39-g340d3c24fc6 8bccc98d839a ef7ec53aa0b8 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:node-exporter.vm07 vm07 *:9100 running (3m) 3s ago 3m 7616k - 1.7.0 72c9c2088986 1859f95972e0 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:node-exporter.vm09 vm09 *:9100 running (2m) 3s ago 2m 7579k - 1.7.0 72c9c2088986 7e9cc4e2059c 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:osd.0 vm09 running (91s) 3s ago 94s 56.4M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 2b2c1a047d82 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:osd.1 vm07 running (91s) 3s ago 95s 36.9M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 5c199f0acc15 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:osd.2 vm07 running (89s) 3s ago 93s 58.5M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a ff80e2683136 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:osd.3 vm09 running (89s) 3s ago 93s 58.0M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 75b3f3ac04da 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:osd.4 vm09 running (87s) 3s ago 91s 56.0M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 527a66f352ad 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:osd.5 vm07 running (87s) 3s ago 91s 60.7M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 5f3f015dd990 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:osd.6 vm07 running (86s) 3s ago 88s 36.7M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a ce94a3b6ab7b 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:osd.7 vm09 running (86s) 3s ago 89s 36.4M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 80f17c672a66 2026-03-07T07:57:03.340 INFO:teuthology.orchestra.run.vm07.stdout:prometheus.vm07 vm07 *:9095 running (2m) 3s ago 2m 36.0M - 2.51.0 1d3b7f56885b 34106be53089 2026-03-07T07:57:03.398 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- bash -c 'ceph orch ls' 2026-03-07T07:57:04.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:04 vm09 bash[21220]: cluster 2026-03-07T06:57:03.101786+0000 mgr.vm07.yrfcuj (mgr.14201) 137 : cluster [DBG] pgmap v90: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:04.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:04 vm09 bash[21220]: cluster 2026-03-07T06:57:03.101786+0000 mgr.vm07.yrfcuj (mgr.14201) 137 : cluster [DBG] pgmap v90: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:04.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:04 vm09 bash[21220]: audit 2026-03-07T06:57:03.329976+0000 mgr.vm07.yrfcuj (mgr.14201) 138 : audit [DBG] from='client.14464 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:04.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:04 vm09 bash[21220]: audit 2026-03-07T06:57:03.329976+0000 mgr.vm07.yrfcuj (mgr.14201) 138 : audit [DBG] from='client.14464 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:04.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:04 vm07 bash[17031]: cluster 2026-03-07T06:57:03.101786+0000 mgr.vm07.yrfcuj (mgr.14201) 137 : cluster [DBG] pgmap v90: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:04.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:04 vm07 bash[17031]: cluster 2026-03-07T06:57:03.101786+0000 mgr.vm07.yrfcuj (mgr.14201) 137 : cluster [DBG] pgmap v90: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:04.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:04 vm07 bash[17031]: audit 2026-03-07T06:57:03.329976+0000 mgr.vm07.yrfcuj (mgr.14201) 138 : audit [DBG] from='client.14464 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:04.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:04 vm07 bash[17031]: audit 2026-03-07T06:57:03.329976+0000 mgr.vm07.yrfcuj (mgr.14201) 138 : audit [DBG] from='client.14464 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:06.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:06 vm09 bash[21220]: cluster 2026-03-07T06:57:05.102074+0000 mgr.vm07.yrfcuj (mgr.14201) 139 : cluster [DBG] pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:06 vm09 bash[21220]: cluster 2026-03-07T06:57:05.102074+0000 mgr.vm07.yrfcuj (mgr.14201) 139 : cluster [DBG] pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:06.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:06 vm07 bash[17031]: cluster 2026-03-07T06:57:05.102074+0000 mgr.vm07.yrfcuj (mgr.14201) 139 : cluster [DBG] pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:06.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:06 vm07 bash[17031]: cluster 2026-03-07T06:57:05.102074+0000 mgr.vm07.yrfcuj (mgr.14201) 139 : cluster [DBG] pgmap v91: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:07.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:07 vm09 bash[21220]: cluster 2026-03-07T06:57:07.102395+0000 mgr.vm07.yrfcuj (mgr.14201) 140 : cluster [DBG] pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:07.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:07 vm09 bash[21220]: cluster 2026-03-07T06:57:07.102395+0000 mgr.vm07.yrfcuj (mgr.14201) 140 : cluster [DBG] pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:07.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:07 vm07 bash[17031]: cluster 2026-03-07T06:57:07.102395+0000 mgr.vm07.yrfcuj (mgr.14201) 140 : cluster [DBG] pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:07.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:07 vm07 bash[17031]: cluster 2026-03-07T06:57:07.102395+0000 mgr.vm07.yrfcuj (mgr.14201) 140 : cluster [DBG] pgmap v92: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:08.192 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:57:08.808 INFO:teuthology.orchestra.run.vm07.stdout:NAME PORTS RUNNING REFRESHED AGE PLACEMENT 2026-03-07T07:57:08.809 INFO:teuthology.orchestra.run.vm07.stdout:alertmanager ?:9093,9094 1/1 9s ago 3m count:1 2026-03-07T07:57:08.809 INFO:teuthology.orchestra.run.vm07.stdout:ceph-exporter 2/2 9s ago 3m * 2026-03-07T07:57:08.809 INFO:teuthology.orchestra.run.vm07.stdout:crash 2/2 9s ago 3m * 2026-03-07T07:57:08.809 INFO:teuthology.orchestra.run.vm07.stdout:grafana ?:3000 1/1 9s ago 3m count:1 2026-03-07T07:57:08.809 INFO:teuthology.orchestra.run.vm07.stdout:mgr 2/2 9s ago 3m count:2 2026-03-07T07:57:08.809 INFO:teuthology.orchestra.run.vm07.stdout:mon 2/2 9s ago 2m vm07:192.168.123.107=vm07;vm09:192.168.123.109=vm09;count:2 2026-03-07T07:57:08.809 INFO:teuthology.orchestra.run.vm07.stdout:node-exporter ?:9100 2/2 9s ago 3m * 2026-03-07T07:57:08.809 INFO:teuthology.orchestra.run.vm07.stdout:osd.all-available-devices 8 9s ago 2m * 2026-03-07T07:57:08.809 INFO:teuthology.orchestra.run.vm07.stdout:prometheus ?:9095 1/1 9s ago 3m count:1 2026-03-07T07:57:08.883 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- bash -c 'ceph orch host ls' 2026-03-07T07:57:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:10 vm09 bash[21220]: audit 2026-03-07T06:57:08.801176+0000 mgr.vm07.yrfcuj (mgr.14201) 141 : audit [DBG] from='client.14468 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:10 vm09 bash[21220]: audit 2026-03-07T06:57:08.801176+0000 mgr.vm07.yrfcuj (mgr.14201) 141 : audit [DBG] from='client.14468 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:10 vm09 bash[21220]: cluster 2026-03-07T06:57:09.105605+0000 mgr.vm07.yrfcuj (mgr.14201) 142 : cluster [DBG] pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:10 vm09 bash[21220]: cluster 2026-03-07T06:57:09.105605+0000 mgr.vm07.yrfcuj (mgr.14201) 142 : cluster [DBG] pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:10 vm09 bash[21220]: audit 2026-03-07T06:57:10.126537+0000 mon.vm07 (mon.0) 651 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:57:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:10 vm09 bash[21220]: audit 2026-03-07T06:57:10.126537+0000 mon.vm07 (mon.0) 651 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:57:10.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:10 vm07 bash[17031]: audit 2026-03-07T06:57:08.801176+0000 mgr.vm07.yrfcuj (mgr.14201) 141 : audit [DBG] from='client.14468 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:10.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:10 vm07 bash[17031]: audit 2026-03-07T06:57:08.801176+0000 mgr.vm07.yrfcuj (mgr.14201) 141 : audit [DBG] from='client.14468 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:10.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:10 vm07 bash[17031]: cluster 2026-03-07T06:57:09.105605+0000 mgr.vm07.yrfcuj (mgr.14201) 142 : cluster [DBG] pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:10.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:10 vm07 bash[17031]: cluster 2026-03-07T06:57:09.105605+0000 mgr.vm07.yrfcuj (mgr.14201) 142 : cluster [DBG] pgmap v93: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:10.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:10 vm07 bash[17031]: audit 2026-03-07T06:57:10.126537+0000 mon.vm07 (mon.0) 651 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:57:10.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:10 vm07 bash[17031]: audit 2026-03-07T06:57:10.126537+0000 mon.vm07 (mon.0) 651 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:57:12.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:12 vm09 bash[21220]: cluster 2026-03-07T06:57:11.105833+0000 mgr.vm07.yrfcuj (mgr.14201) 143 : cluster [DBG] pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:12.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:12 vm09 bash[21220]: cluster 2026-03-07T06:57:11.105833+0000 mgr.vm07.yrfcuj (mgr.14201) 143 : cluster [DBG] pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:12.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:12 vm07 bash[17031]: cluster 2026-03-07T06:57:11.105833+0000 mgr.vm07.yrfcuj (mgr.14201) 143 : cluster [DBG] pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:12.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:12 vm07 bash[17031]: cluster 2026-03-07T06:57:11.105833+0000 mgr.vm07.yrfcuj (mgr.14201) 143 : cluster [DBG] pgmap v94: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:13.649 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:57:14.017 INFO:teuthology.orchestra.run.vm07.stdout:HOST ADDR LABELS STATUS 2026-03-07T07:57:14.018 INFO:teuthology.orchestra.run.vm07.stdout:vm07 192.168.123.107 2026-03-07T07:57:14.018 INFO:teuthology.orchestra.run.vm07.stdout:vm09 192.168.123.109 2026-03-07T07:57:14.018 INFO:teuthology.orchestra.run.vm07.stdout:2 hosts in cluster 2026-03-07T07:57:14.075 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- bash -c 'ceph orch device ls' 2026-03-07T07:57:14.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:14 vm07 bash[17031]: cluster 2026-03-07T06:57:13.106080+0000 mgr.vm07.yrfcuj (mgr.14201) 144 : cluster [DBG] pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:14.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:14 vm07 bash[17031]: cluster 2026-03-07T06:57:13.106080+0000 mgr.vm07.yrfcuj (mgr.14201) 144 : cluster [DBG] pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:14.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:14 vm09 bash[21220]: cluster 2026-03-07T06:57:13.106080+0000 mgr.vm07.yrfcuj (mgr.14201) 144 : cluster [DBG] pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:14.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:14 vm09 bash[21220]: cluster 2026-03-07T06:57:13.106080+0000 mgr.vm07.yrfcuj (mgr.14201) 144 : cluster [DBG] pgmap v95: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:15.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:15 vm09 bash[21220]: audit 2026-03-07T06:57:14.011852+0000 mgr.vm07.yrfcuj (mgr.14201) 145 : audit [DBG] from='client.24305 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:15.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:15 vm09 bash[21220]: audit 2026-03-07T06:57:14.011852+0000 mgr.vm07.yrfcuj (mgr.14201) 145 : audit [DBG] from='client.24305 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:15.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:15 vm07 bash[17031]: audit 2026-03-07T06:57:14.011852+0000 mgr.vm07.yrfcuj (mgr.14201) 145 : audit [DBG] from='client.24305 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:15.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:15 vm07 bash[17031]: audit 2026-03-07T06:57:14.011852+0000 mgr.vm07.yrfcuj (mgr.14201) 145 : audit [DBG] from='client.24305 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:16.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:16 vm09 bash[21220]: cluster 2026-03-07T06:57:15.106314+0000 mgr.vm07.yrfcuj (mgr.14201) 146 : cluster [DBG] pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:16 vm09 bash[21220]: cluster 2026-03-07T06:57:15.106314+0000 mgr.vm07.yrfcuj (mgr.14201) 146 : cluster [DBG] pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:16.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:16 vm07 bash[17031]: cluster 2026-03-07T06:57:15.106314+0000 mgr.vm07.yrfcuj (mgr.14201) 146 : cluster [DBG] pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:16.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:16 vm07 bash[17031]: cluster 2026-03-07T06:57:15.106314+0000 mgr.vm07.yrfcuj (mgr.14201) 146 : cluster [DBG] pgmap v96: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:18.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:18 vm09 bash[21220]: cluster 2026-03-07T06:57:17.106586+0000 mgr.vm07.yrfcuj (mgr.14201) 147 : cluster [DBG] pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:18.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:18 vm09 bash[21220]: cluster 2026-03-07T06:57:17.106586+0000 mgr.vm07.yrfcuj (mgr.14201) 147 : cluster [DBG] pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:18 vm07 bash[17031]: cluster 2026-03-07T06:57:17.106586+0000 mgr.vm07.yrfcuj (mgr.14201) 147 : cluster [DBG] pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:18.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:18 vm07 bash[17031]: cluster 2026-03-07T06:57:17.106586+0000 mgr.vm07.yrfcuj (mgr.14201) 147 : cluster [DBG] pgmap v97: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:18.879 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:57:19.228 INFO:teuthology.orchestra.run.vm07.stdout:HOST PATH TYPE DEVICE ID SIZE AVAILABLE REFRESHED REJECT REASONS 2026-03-07T07:57:19.228 INFO:teuthology.orchestra.run.vm07.stdout:vm07 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 85s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-07T07:57:19.228 INFO:teuthology.orchestra.run.vm07.stdout:vm07 /dev/vdb hdd DWNBRSTVMM07001 20.0G No 85s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T07:57:19.228 INFO:teuthology.orchestra.run.vm07.stdout:vm07 /dev/vdc hdd DWNBRSTVMM07002 20.0G No 85s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T07:57:19.228 INFO:teuthology.orchestra.run.vm07.stdout:vm07 /dev/vdd hdd DWNBRSTVMM07003 20.0G No 85s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T07:57:19.228 INFO:teuthology.orchestra.run.vm07.stdout:vm07 /dev/vde hdd DWNBRSTVMM07004 20.0G No 85s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T07:57:19.228 INFO:teuthology.orchestra.run.vm07.stdout:vm09 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 84s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-07T07:57:19.228 INFO:teuthology.orchestra.run.vm07.stdout:vm09 /dev/vdb hdd DWNBRSTVMM09001 20.0G No 84s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T07:57:19.228 INFO:teuthology.orchestra.run.vm07.stdout:vm09 /dev/vdc hdd DWNBRSTVMM09002 20.0G No 84s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T07:57:19.228 INFO:teuthology.orchestra.run.vm07.stdout:vm09 /dev/vdd hdd DWNBRSTVMM09003 20.0G No 84s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T07:57:19.228 INFO:teuthology.orchestra.run.vm07.stdout:vm09 /dev/vde hdd DWNBRSTVMM09004 20.0G No 84s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T07:57:19.292 INFO:teuthology.run_tasks:Running task vip... 2026-03-07T07:57:19.295 INFO:tasks.vip:Allocating static IPs for each host... 2026-03-07T07:57:19.295 INFO:tasks.vip:peername 192.168.123.107 2026-03-07T07:57:19.295 INFO:tasks.vip:192.168.123.107 in 192.168.123.0/24, pos 106 2026-03-07T07:57:19.295 INFO:tasks.vip:vm07.local static 12.12.0.107, vnet 12.12.0.0/22 2026-03-07T07:57:19.295 INFO:tasks.vip:VIPs are [IPv4Address('12.12.1.107')] 2026-03-07T07:57:19.295 DEBUG:teuthology.orchestra.run.vm07:> sudo ip route ls 2026-03-07T07:57:19.302 INFO:teuthology.orchestra.run.vm07.stdout:default via 192.168.123.1 dev ens3 proto dhcp src 192.168.123.107 metric 100 2026-03-07T07:57:19.302 INFO:teuthology.orchestra.run.vm07.stdout:172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.0.1 linkdown 2026-03-07T07:57:19.302 INFO:teuthology.orchestra.run.vm07.stdout:192.168.123.0/24 dev ens3 proto kernel scope link src 192.168.123.107 metric 100 2026-03-07T07:57:19.302 INFO:teuthology.orchestra.run.vm07.stdout:192.168.123.1 dev ens3 proto dhcp scope link src 192.168.123.107 metric 100 2026-03-07T07:57:19.303 INFO:tasks.vip:Configuring 12.12.0.107 on vm07.local iface ens3... 2026-03-07T07:57:19.303 DEBUG:teuthology.orchestra.run.vm07:> sudo ip addr add 12.12.0.107/22 dev ens3 2026-03-07T07:57:19.351 INFO:tasks.vip:peername 192.168.123.109 2026-03-07T07:57:19.351 INFO:tasks.vip:192.168.123.109 in 192.168.123.0/24, pos 108 2026-03-07T07:57:19.351 INFO:tasks.vip:vm09.local static 12.12.0.109, vnet 12.12.0.0/22 2026-03-07T07:57:19.351 DEBUG:teuthology.orchestra.run.vm09:> sudo ip route ls 2026-03-07T07:57:19.358 INFO:teuthology.orchestra.run.vm09.stdout:default via 192.168.123.1 dev ens3 proto dhcp src 192.168.123.109 metric 100 2026-03-07T07:57:19.358 INFO:teuthology.orchestra.run.vm09.stdout:172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.0.1 linkdown 2026-03-07T07:57:19.358 INFO:teuthology.orchestra.run.vm09.stdout:192.168.123.0/24 dev ens3 proto kernel scope link src 192.168.123.109 metric 100 2026-03-07T07:57:19.358 INFO:teuthology.orchestra.run.vm09.stdout:192.168.123.1 dev ens3 proto dhcp scope link src 192.168.123.109 metric 100 2026-03-07T07:57:19.359 INFO:tasks.vip:Configuring 12.12.0.109 on vm09.local iface ens3... 2026-03-07T07:57:19.359 DEBUG:teuthology.orchestra.run.vm09:> sudo ip addr add 12.12.0.109/22 dev ens3 2026-03-07T07:57:19.410 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-07T07:57:19.412 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm07.local 2026-03-07T07:57:19.412 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- bash -c 'ceph orch device ls --refresh' 2026-03-07T07:57:20.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:20 vm09 bash[21220]: cluster 2026-03-07T06:57:19.106855+0000 mgr.vm07.yrfcuj (mgr.14201) 148 : cluster [DBG] pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:20.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:20 vm09 bash[21220]: cluster 2026-03-07T06:57:19.106855+0000 mgr.vm07.yrfcuj (mgr.14201) 148 : cluster [DBG] pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:20.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:20 vm09 bash[21220]: audit 2026-03-07T06:57:19.221356+0000 mgr.vm07.yrfcuj (mgr.14201) 149 : audit [DBG] from='client.24307 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:20.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:20 vm09 bash[21220]: audit 2026-03-07T06:57:19.221356+0000 mgr.vm07.yrfcuj (mgr.14201) 149 : audit [DBG] from='client.24307 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:20.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:20 vm07 bash[17031]: cluster 2026-03-07T06:57:19.106855+0000 mgr.vm07.yrfcuj (mgr.14201) 148 : cluster [DBG] pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:20.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:20 vm07 bash[17031]: cluster 2026-03-07T06:57:19.106855+0000 mgr.vm07.yrfcuj (mgr.14201) 148 : cluster [DBG] pgmap v98: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:20.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:20 vm07 bash[17031]: audit 2026-03-07T06:57:19.221356+0000 mgr.vm07.yrfcuj (mgr.14201) 149 : audit [DBG] from='client.24307 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:20.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:20 vm07 bash[17031]: audit 2026-03-07T06:57:19.221356+0000 mgr.vm07.yrfcuj (mgr.14201) 149 : audit [DBG] from='client.24307 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:22.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:22 vm09 bash[21220]: cluster 2026-03-07T06:57:21.107088+0000 mgr.vm07.yrfcuj (mgr.14201) 150 : cluster [DBG] pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:22.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:22 vm09 bash[21220]: cluster 2026-03-07T06:57:21.107088+0000 mgr.vm07.yrfcuj (mgr.14201) 150 : cluster [DBG] pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:22.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:22 vm07 bash[17031]: cluster 2026-03-07T06:57:21.107088+0000 mgr.vm07.yrfcuj (mgr.14201) 150 : cluster [DBG] pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:22.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:22 vm07 bash[17031]: cluster 2026-03-07T06:57:21.107088+0000 mgr.vm07.yrfcuj (mgr.14201) 150 : cluster [DBG] pgmap v99: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:24.186 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:57:24.542 INFO:teuthology.orchestra.run.vm07.stdout:HOST PATH TYPE DEVICE ID SIZE AVAILABLE REFRESHED REJECT REASONS 2026-03-07T07:57:24.542 INFO:teuthology.orchestra.run.vm07.stdout:vm07 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 90s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-07T07:57:24.542 INFO:teuthology.orchestra.run.vm07.stdout:vm07 /dev/vdb hdd DWNBRSTVMM07001 20.0G No 90s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T07:57:24.542 INFO:teuthology.orchestra.run.vm07.stdout:vm07 /dev/vdc hdd DWNBRSTVMM07002 20.0G No 90s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T07:57:24.542 INFO:teuthology.orchestra.run.vm07.stdout:vm07 /dev/vdd hdd DWNBRSTVMM07003 20.0G No 90s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T07:57:24.542 INFO:teuthology.orchestra.run.vm07.stdout:vm07 /dev/vde hdd DWNBRSTVMM07004 20.0G No 90s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T07:57:24.542 INFO:teuthology.orchestra.run.vm07.stdout:vm09 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 89s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-07T07:57:24.542 INFO:teuthology.orchestra.run.vm07.stdout:vm09 /dev/vdb hdd DWNBRSTVMM09001 20.0G No 89s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T07:57:24.542 INFO:teuthology.orchestra.run.vm07.stdout:vm09 /dev/vdc hdd DWNBRSTVMM09002 20.0G No 89s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T07:57:24.542 INFO:teuthology.orchestra.run.vm07.stdout:vm09 /dev/vdd hdd DWNBRSTVMM09003 20.0G No 89s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T07:57:24.542 INFO:teuthology.orchestra.run.vm07.stdout:vm09 /dev/vde hdd DWNBRSTVMM09004 20.0G No 89s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T07:57:24.569 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:24 vm07 bash[17031]: cluster 2026-03-07T06:57:23.107319+0000 mgr.vm07.yrfcuj (mgr.14201) 151 : cluster [DBG] pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:24.569 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:24 vm07 bash[17031]: cluster 2026-03-07T06:57:23.107319+0000 mgr.vm07.yrfcuj (mgr.14201) 151 : cluster [DBG] pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:24.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:24 vm09 bash[21220]: cluster 2026-03-07T06:57:23.107319+0000 mgr.vm07.yrfcuj (mgr.14201) 151 : cluster [DBG] pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:24.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:24 vm09 bash[21220]: cluster 2026-03-07T06:57:23.107319+0000 mgr.vm07.yrfcuj (mgr.14201) 151 : cluster [DBG] pgmap v100: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:24.631 INFO:teuthology.run_tasks:Running task cephadm.apply... 2026-03-07T07:57:24.635 INFO:tasks.cephadm:Applying spec(s): placement: count: 4 host_pattern: '*' service_id: foo service_type: rgw spec: rgw_frontend_port: 8000 --- placement: count: 2 service_id: rgw.foo service_type: ingress spec: backend_service: rgw.foo frontend_port: 9000 monitor_port: 9001 virtual_ip: 12.12.1.107/22 2026-03-07T07:57:24.635 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph orch apply -i - 2026-03-07T07:57:25.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:25 vm09 bash[21220]: audit 2026-03-07T06:57:24.530639+0000 mon.vm07 (mon.0) 652 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:57:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:25 vm09 bash[21220]: audit 2026-03-07T06:57:24.530639+0000 mon.vm07 (mon.0) 652 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:57:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:25 vm09 bash[21220]: audit 2026-03-07T06:57:25.126910+0000 mon.vm07 (mon.0) 653 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:57:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:25 vm09 bash[21220]: audit 2026-03-07T06:57:25.126910+0000 mon.vm07 (mon.0) 653 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:57:25.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:25 vm07 bash[17031]: audit 2026-03-07T06:57:24.530639+0000 mon.vm07 (mon.0) 652 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:57:25.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:25 vm07 bash[17031]: audit 2026-03-07T06:57:24.530639+0000 mon.vm07 (mon.0) 652 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:57:25.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:25 vm07 bash[17031]: audit 2026-03-07T06:57:25.126910+0000 mon.vm07 (mon.0) 653 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:57:25.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:25 vm07 bash[17031]: audit 2026-03-07T06:57:25.126910+0000 mon.vm07 (mon.0) 653 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:57:26.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:26 vm09 bash[21220]: audit 2026-03-07T06:57:24.530017+0000 mgr.vm07.yrfcuj (mgr.14201) 152 : audit [DBG] from='client.14480 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:26.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:26 vm09 bash[21220]: audit 2026-03-07T06:57:24.530017+0000 mgr.vm07.yrfcuj (mgr.14201) 152 : audit [DBG] from='client.14480 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:26.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:26 vm09 bash[21220]: cluster 2026-03-07T06:57:25.107594+0000 mgr.vm07.yrfcuj (mgr.14201) 153 : cluster [DBG] pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:26.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:26 vm09 bash[21220]: cluster 2026-03-07T06:57:25.107594+0000 mgr.vm07.yrfcuj (mgr.14201) 153 : cluster [DBG] pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:26.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:26 vm07 bash[17031]: audit 2026-03-07T06:57:24.530017+0000 mgr.vm07.yrfcuj (mgr.14201) 152 : audit [DBG] from='client.14480 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:26.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:26 vm07 bash[17031]: audit 2026-03-07T06:57:24.530017+0000 mgr.vm07.yrfcuj (mgr.14201) 152 : audit [DBG] from='client.14480 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:26.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:26 vm07 bash[17031]: cluster 2026-03-07T06:57:25.107594+0000 mgr.vm07.yrfcuj (mgr.14201) 153 : cluster [DBG] pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:26.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:26 vm07 bash[17031]: cluster 2026-03-07T06:57:25.107594+0000 mgr.vm07.yrfcuj (mgr.14201) 153 : cluster [DBG] pgmap v101: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:28.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:28 vm09 bash[21220]: cluster 2026-03-07T06:57:27.107869+0000 mgr.vm07.yrfcuj (mgr.14201) 154 : cluster [DBG] pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:28.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:28 vm09 bash[21220]: cluster 2026-03-07T06:57:27.107869+0000 mgr.vm07.yrfcuj (mgr.14201) 154 : cluster [DBG] pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:28.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:28 vm07 bash[17031]: cluster 2026-03-07T06:57:27.107869+0000 mgr.vm07.yrfcuj (mgr.14201) 154 : cluster [DBG] pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:28.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:28 vm07 bash[17031]: cluster 2026-03-07T06:57:27.107869+0000 mgr.vm07.yrfcuj (mgr.14201) 154 : cluster [DBG] pgmap v102: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:29.351 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:57:29.857 INFO:teuthology.orchestra.run.vm07.stdout:Scheduled rgw.foo update... 2026-03-07T07:57:29.857 INFO:teuthology.orchestra.run.vm07.stdout:Scheduled ingress.rgw.foo update... 2026-03-07T07:57:29.958 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-07T07:57:29.960 INFO:tasks.cephadm:Waiting for ceph service rgw.foo to start (timeout 300)... 2026-03-07T07:57:29.960 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph orch ls -f json 2026-03-07T07:57:30.235 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:30 vm07 bash[17031]: cluster 2026-03-07T06:57:29.108170+0000 mgr.vm07.yrfcuj (mgr.14201) 155 : cluster [DBG] pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:30.235 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:30 vm07 bash[17031]: cluster 2026-03-07T06:57:29.108170+0000 mgr.vm07.yrfcuj (mgr.14201) 155 : cluster [DBG] pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:30.235 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:30 vm07 bash[17031]: audit 2026-03-07T06:57:29.603993+0000 mon.vm07 (mon.0) 654 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.235 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:30 vm07 bash[17031]: audit 2026-03-07T06:57:29.603993+0000 mon.vm07 (mon.0) 654 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.235 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:30 vm07 bash[17031]: audit 2026-03-07T06:57:29.622435+0000 mon.vm07 (mon.0) 655 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.235 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:30 vm07 bash[17031]: audit 2026-03-07T06:57:29.622435+0000 mon.vm07 (mon.0) 655 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.235 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:30 vm07 bash[17031]: audit 2026-03-07T06:57:29.845432+0000 mon.vm07 (mon.0) 656 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.236 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:30 vm07 bash[17031]: audit 2026-03-07T06:57:29.845432+0000 mon.vm07 (mon.0) 656 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.236 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:30 vm07 bash[17031]: audit 2026-03-07T06:57:29.850222+0000 mon.vm07 (mon.0) 657 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.236 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:30 vm07 bash[17031]: audit 2026-03-07T06:57:29.850222+0000 mon.vm07 (mon.0) 657 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.236 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:30 vm07 bash[17031]: audit 2026-03-07T06:57:29.887738+0000 mon.vm07 (mon.0) 658 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.236 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:30 vm07 bash[17031]: audit 2026-03-07T06:57:29.887738+0000 mon.vm07 (mon.0) 658 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.236 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:30 vm07 bash[17031]: audit 2026-03-07T06:57:29.892267+0000 mon.vm07 (mon.0) 659 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.236 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:30 vm07 bash[17031]: audit 2026-03-07T06:57:29.892267+0000 mon.vm07 (mon.0) 659 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.236 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:30 vm07 bash[17031]: audit 2026-03-07T06:57:30.072374+0000 mon.vm07 (mon.0) 660 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.236 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:30 vm07 bash[17031]: audit 2026-03-07T06:57:30.072374+0000 mon.vm07 (mon.0) 660 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.236 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:30 vm07 bash[17031]: audit 2026-03-07T06:57:30.077022+0000 mon.vm07 (mon.0) 661 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.236 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:30 vm07 bash[17031]: audit 2026-03-07T06:57:30.077022+0000 mon.vm07 (mon.0) 661 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:30 vm09 bash[21220]: cluster 2026-03-07T06:57:29.108170+0000 mgr.vm07.yrfcuj (mgr.14201) 155 : cluster [DBG] pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:30.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:30 vm09 bash[21220]: cluster 2026-03-07T06:57:29.108170+0000 mgr.vm07.yrfcuj (mgr.14201) 155 : cluster [DBG] pgmap v103: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:30.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:30 vm09 bash[21220]: audit 2026-03-07T06:57:29.603993+0000 mon.vm07 (mon.0) 654 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:30 vm09 bash[21220]: audit 2026-03-07T06:57:29.603993+0000 mon.vm07 (mon.0) 654 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:30 vm09 bash[21220]: audit 2026-03-07T06:57:29.622435+0000 mon.vm07 (mon.0) 655 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:30 vm09 bash[21220]: audit 2026-03-07T06:57:29.622435+0000 mon.vm07 (mon.0) 655 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:30 vm09 bash[21220]: audit 2026-03-07T06:57:29.845432+0000 mon.vm07 (mon.0) 656 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:30 vm09 bash[21220]: audit 2026-03-07T06:57:29.845432+0000 mon.vm07 (mon.0) 656 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:30 vm09 bash[21220]: audit 2026-03-07T06:57:29.850222+0000 mon.vm07 (mon.0) 657 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:30 vm09 bash[21220]: audit 2026-03-07T06:57:29.850222+0000 mon.vm07 (mon.0) 657 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:30 vm09 bash[21220]: audit 2026-03-07T06:57:29.887738+0000 mon.vm07 (mon.0) 658 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:30 vm09 bash[21220]: audit 2026-03-07T06:57:29.887738+0000 mon.vm07 (mon.0) 658 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:30 vm09 bash[21220]: audit 2026-03-07T06:57:29.892267+0000 mon.vm07 (mon.0) 659 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:30 vm09 bash[21220]: audit 2026-03-07T06:57:29.892267+0000 mon.vm07 (mon.0) 659 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:30 vm09 bash[21220]: audit 2026-03-07T06:57:30.072374+0000 mon.vm07 (mon.0) 660 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:30 vm09 bash[21220]: audit 2026-03-07T06:57:30.072374+0000 mon.vm07 (mon.0) 660 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:30 vm09 bash[21220]: audit 2026-03-07T06:57:30.077022+0000 mon.vm07 (mon.0) 661 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:30.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:30 vm09 bash[21220]: audit 2026-03-07T06:57:30.077022+0000 mon.vm07 (mon.0) 661 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:31.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:31 vm09 bash[21220]: audit 2026-03-07T06:57:29.839447+0000 mgr.vm07.yrfcuj (mgr.14201) 156 : audit [DBG] from='client.14484 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:31.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:31 vm09 bash[21220]: audit 2026-03-07T06:57:29.839447+0000 mgr.vm07.yrfcuj (mgr.14201) 156 : audit [DBG] from='client.14484 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:31.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:31 vm09 bash[21220]: cephadm 2026-03-07T06:57:29.841494+0000 mgr.vm07.yrfcuj (mgr.14201) 157 : cephadm [INF] Saving service rgw.foo spec with placement count:4;* 2026-03-07T07:57:31.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:31 vm09 bash[21220]: cephadm 2026-03-07T06:57:29.841494+0000 mgr.vm07.yrfcuj (mgr.14201) 157 : cephadm [INF] Saving service rgw.foo spec with placement count:4;* 2026-03-07T07:57:31.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:31 vm09 bash[21220]: cephadm 2026-03-07T06:57:29.846003+0000 mgr.vm07.yrfcuj (mgr.14201) 158 : cephadm [INF] Saving service ingress.rgw.foo spec with placement count:2 2026-03-07T07:57:31.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:31 vm09 bash[21220]: cephadm 2026-03-07T06:57:29.846003+0000 mgr.vm07.yrfcuj (mgr.14201) 158 : cephadm [INF] Saving service ingress.rgw.foo spec with placement count:2 2026-03-07T07:57:31.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:31 vm09 bash[21220]: audit 2026-03-07T06:57:30.339187+0000 mon.vm07 (mon.0) 662 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:31.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:31 vm09 bash[21220]: audit 2026-03-07T06:57:30.339187+0000 mon.vm07 (mon.0) 662 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:31.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:31 vm09 bash[21220]: audit 2026-03-07T06:57:30.343286+0000 mon.vm07 (mon.0) 663 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:31.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:31 vm09 bash[21220]: audit 2026-03-07T06:57:30.343286+0000 mon.vm07 (mon.0) 663 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:31.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:31 vm07 bash[17031]: audit 2026-03-07T06:57:29.839447+0000 mgr.vm07.yrfcuj (mgr.14201) 156 : audit [DBG] from='client.14484 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:31.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:31 vm07 bash[17031]: audit 2026-03-07T06:57:29.839447+0000 mgr.vm07.yrfcuj (mgr.14201) 156 : audit [DBG] from='client.14484 -' entity='client.admin' cmd=[{"prefix": "orch apply", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:57:31.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:31 vm07 bash[17031]: cephadm 2026-03-07T06:57:29.841494+0000 mgr.vm07.yrfcuj (mgr.14201) 157 : cephadm [INF] Saving service rgw.foo spec with placement count:4;* 2026-03-07T07:57:31.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:31 vm07 bash[17031]: cephadm 2026-03-07T06:57:29.841494+0000 mgr.vm07.yrfcuj (mgr.14201) 157 : cephadm [INF] Saving service rgw.foo spec with placement count:4;* 2026-03-07T07:57:31.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:31 vm07 bash[17031]: cephadm 2026-03-07T06:57:29.846003+0000 mgr.vm07.yrfcuj (mgr.14201) 158 : cephadm [INF] Saving service ingress.rgw.foo spec with placement count:2 2026-03-07T07:57:31.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:31 vm07 bash[17031]: cephadm 2026-03-07T06:57:29.846003+0000 mgr.vm07.yrfcuj (mgr.14201) 158 : cephadm [INF] Saving service ingress.rgw.foo spec with placement count:2 2026-03-07T07:57:31.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:31 vm07 bash[17031]: audit 2026-03-07T06:57:30.339187+0000 mon.vm07 (mon.0) 662 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:31.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:31 vm07 bash[17031]: audit 2026-03-07T06:57:30.339187+0000 mon.vm07 (mon.0) 662 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:31.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:31 vm07 bash[17031]: audit 2026-03-07T06:57:30.343286+0000 mon.vm07 (mon.0) 663 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:31.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:31 vm07 bash[17031]: audit 2026-03-07T06:57:30.343286+0000 mon.vm07 (mon.0) 663 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:32.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:32 vm09 bash[21220]: cluster 2026-03-07T06:57:31.108468+0000 mgr.vm07.yrfcuj (mgr.14201) 159 : cluster [DBG] pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:32.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:32 vm09 bash[21220]: cluster 2026-03-07T06:57:31.108468+0000 mgr.vm07.yrfcuj (mgr.14201) 159 : cluster [DBG] pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:32.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:32 vm07 bash[17031]: cluster 2026-03-07T06:57:31.108468+0000 mgr.vm07.yrfcuj (mgr.14201) 159 : cluster [DBG] pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:32.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:32 vm07 bash[17031]: cluster 2026-03-07T06:57:31.108468+0000 mgr.vm07.yrfcuj (mgr.14201) 159 : cluster [DBG] pgmap v104: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:34.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:34 vm09 bash[21220]: cluster 2026-03-07T06:57:33.108772+0000 mgr.vm07.yrfcuj (mgr.14201) 160 : cluster [DBG] pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:34.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:34 vm09 bash[21220]: cluster 2026-03-07T06:57:33.108772+0000 mgr.vm07.yrfcuj (mgr.14201) 160 : cluster [DBG] pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:34.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:34 vm07 bash[17031]: cluster 2026-03-07T06:57:33.108772+0000 mgr.vm07.yrfcuj (mgr.14201) 160 : cluster [DBG] pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:34.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:34 vm07 bash[17031]: cluster 2026-03-07T06:57:33.108772+0000 mgr.vm07.yrfcuj (mgr.14201) 160 : cluster [DBG] pgmap v105: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:34.785 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:57:35.204 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:57:35.204 INFO:teuthology.orchestra.run.vm07.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-07T06:53:25.695727Z", "last_refresh": "2026-03-07T06:57:30.067219Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-07T06:54:32.927755Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-07T06:53:24.247577Z", "last_refresh": "2026-03-07T06:57:29.595811Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:33.727624Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-07T06:53:23.837568Z", "last_refresh": "2026-03-07T06:57:29.595876Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-07T06:53:24.987960Z", "last_refresh": "2026-03-07T06:57:30.067033Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-07T06:57:29.850523Z service:ingress.rgw.foo [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "rgw.foo", "service_name": "ingress.rgw.foo", "service_type": "ingress", "spec": {"backend_service": "rgw.foo", "first_virtual_router_id": 50, "frontend_port": 9000, "monitor_port": 9001, "virtual_ip": "12.12.1.107/22"}, "status": {"created": "2026-03-07T06:57:29.846008Z", "ports": [9000, 9001], "running": 0, "size": 4, "virtual_ip": "12.12.1.107/22"}}, {"events": ["2026-03-07T06:54:35.185395Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-07T06:53:23.396871Z", "last_refresh": "2026-03-07T06:57:29.595909Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:36.226821Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm07:192.168.123.107=vm07", "vm09:192.168.123.109=vm09"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-07T06:54:13.493474Z", "last_refresh": "2026-03-07T06:57:29.595844Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:34.440095Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-07T06:53:25.333423Z", "last_refresh": "2026-03-07T06:57:29.596000Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:54.121213Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-07T06:54:54.116553Z", "last_refresh": "2026-03-07T06:57:29.595721Z", "running": 8, "size": 8}}, {"events": ["2026-03-07T06:54:36.229563Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-07T06:53:24.610265Z", "last_refresh": "2026-03-07T06:57:30.066979Z", "ports": [9095], "running": 1, "size": 1}}, {"events": ["2026-03-07T06:57:29.845838Z service:rgw.foo [INFO] \"service was created\""], "placement": {"count": 4, "host_pattern": "*"}, "service_id": "foo", "service_name": "rgw.foo", "service_type": "rgw", "spec": {"rgw_frontend_port": 8000}, "status": {"created": "2026-03-07T06:57:29.841506Z", "ports": [8000], "running": 0, "size": 4}}] 2026-03-07T07:57:35.272 INFO:tasks.cephadm:rgw.foo has 0/4 2026-03-07T07:57:36.272 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph orch ls -f json 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: cluster 2026-03-07T06:57:35.109043+0000 mgr.vm07.yrfcuj (mgr.14201) 161 : cluster [DBG] pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: cluster 2026-03-07T06:57:35.109043+0000 mgr.vm07.yrfcuj (mgr.14201) 161 : cluster [DBG] pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:35.197454+0000 mgr.vm07.yrfcuj (mgr.14201) 162 : audit [DBG] from='client.14488 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:35.197454+0000 mgr.vm07.yrfcuj (mgr.14201) 162 : audit [DBG] from='client.14488 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:35.980504+0000 mon.vm07 (mon.0) 664 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:35.980504+0000 mon.vm07 (mon.0) 664 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:35.984180+0000 mon.vm07 (mon.0) 665 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:35.984180+0000 mon.vm07 (mon.0) 665 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.176516+0000 mon.vm07 (mon.0) 666 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.176516+0000 mon.vm07 (mon.0) 666 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.180732+0000 mon.vm07 (mon.0) 667 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.180732+0000 mon.vm07 (mon.0) 667 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.181430+0000 mon.vm07 (mon.0) 668 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.181430+0000 mon.vm07 (mon.0) 668 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.181849+0000 mon.vm07 (mon.0) 669 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.181849+0000 mon.vm07 (mon.0) 669 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.184859+0000 mon.vm07 (mon.0) 670 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.184859+0000 mon.vm07 (mon.0) 670 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.186090+0000 mon.vm07 (mon.0) 671 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.186090+0000 mon.vm07 (mon.0) 671 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.188699+0000 mon.vm07 (mon.0) 672 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm09.fgzozy", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.188699+0000 mon.vm07 (mon.0) 672 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm09.fgzozy", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.190602+0000 mon.vm07 (mon.0) 673 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm09.fgzozy", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.190602+0000 mon.vm07 (mon.0) 673 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm09.fgzozy", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.193784+0000 mon.vm07 (mon.0) 674 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.193784+0000 mon.vm07 (mon.0) 674 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.194682+0000 mon.vm07 (mon.0) 675 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:36.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:36 vm07 bash[17031]: audit 2026-03-07T06:57:36.194682+0000 mon.vm07 (mon.0) 675 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: cluster 2026-03-07T06:57:35.109043+0000 mgr.vm07.yrfcuj (mgr.14201) 161 : cluster [DBG] pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: cluster 2026-03-07T06:57:35.109043+0000 mgr.vm07.yrfcuj (mgr.14201) 161 : cluster [DBG] pgmap v106: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:35.197454+0000 mgr.vm07.yrfcuj (mgr.14201) 162 : audit [DBG] from='client.14488 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:35.197454+0000 mgr.vm07.yrfcuj (mgr.14201) 162 : audit [DBG] from='client.14488 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:35.980504+0000 mon.vm07 (mon.0) 664 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:35.980504+0000 mon.vm07 (mon.0) 664 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:35.984180+0000 mon.vm07 (mon.0) 665 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:35.984180+0000 mon.vm07 (mon.0) 665 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.176516+0000 mon.vm07 (mon.0) 666 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.176516+0000 mon.vm07 (mon.0) 666 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.180732+0000 mon.vm07 (mon.0) 667 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.180732+0000 mon.vm07 (mon.0) 667 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.181430+0000 mon.vm07 (mon.0) 668 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.181430+0000 mon.vm07 (mon.0) 668 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.181849+0000 mon.vm07 (mon.0) 669 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.181849+0000 mon.vm07 (mon.0) 669 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.184859+0000 mon.vm07 (mon.0) 670 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.184859+0000 mon.vm07 (mon.0) 670 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.186090+0000 mon.vm07 (mon.0) 671 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.186090+0000 mon.vm07 (mon.0) 671 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.188699+0000 mon.vm07 (mon.0) 672 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm09.fgzozy", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.188699+0000 mon.vm07 (mon.0) 672 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm09.fgzozy", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.190602+0000 mon.vm07 (mon.0) 673 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm09.fgzozy", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.190602+0000 mon.vm07 (mon.0) 673 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm09.fgzozy", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.193784+0000 mon.vm07 (mon.0) 674 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.193784+0000 mon.vm07 (mon.0) 674 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:36.735 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.194682+0000 mon.vm07 (mon.0) 675 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:36.736 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 bash[21220]: audit 2026-03-07T06:57:36.194682+0000 mon.vm07 (mon.0) 675 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:36.985 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:57:37.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:36 vm09 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:57:37.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:37 vm07 bash[17031]: cephadm 2026-03-07T06:57:36.195071+0000 mgr.vm07.yrfcuj (mgr.14201) 163 : cephadm [INF] Deploying daemon rgw.foo.vm09.fgzozy on vm09 2026-03-07T07:57:37.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:37 vm07 bash[17031]: cephadm 2026-03-07T06:57:36.195071+0000 mgr.vm07.yrfcuj (mgr.14201) 163 : cephadm [INF] Deploying daemon rgw.foo.vm09.fgzozy on vm09 2026-03-07T07:57:37.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:37 vm07 bash[17031]: audit 2026-03-07T06:57:37.071313+0000 mon.vm07 (mon.0) 676 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:37.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:37 vm07 bash[17031]: audit 2026-03-07T06:57:37.071313+0000 mon.vm07 (mon.0) 676 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:37.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:37 vm07 bash[17031]: audit 2026-03-07T06:57:37.075757+0000 mon.vm07 (mon.0) 677 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:37.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:37 vm07 bash[17031]: audit 2026-03-07T06:57:37.075757+0000 mon.vm07 (mon.0) 677 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:37.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:37 vm07 bash[17031]: audit 2026-03-07T06:57:37.079856+0000 mon.vm07 (mon.0) 678 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:37.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:37 vm07 bash[17031]: audit 2026-03-07T06:57:37.079856+0000 mon.vm07 (mon.0) 678 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:37.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:37 vm07 bash[17031]: audit 2026-03-07T06:57:37.080317+0000 mon.vm07 (mon.0) 679 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm07.bgfaka", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-07T07:57:37.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:37 vm07 bash[17031]: audit 2026-03-07T06:57:37.080317+0000 mon.vm07 (mon.0) 679 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm07.bgfaka", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-07T07:57:37.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:37 vm07 bash[17031]: audit 2026-03-07T06:57:37.087888+0000 mon.vm07 (mon.0) 680 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm07.bgfaka", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-07T07:57:37.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:37 vm07 bash[17031]: audit 2026-03-07T06:57:37.087888+0000 mon.vm07 (mon.0) 680 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm07.bgfaka", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-07T07:57:37.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:37 vm07 bash[17031]: audit 2026-03-07T06:57:37.092281+0000 mon.vm07 (mon.0) 681 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:37.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:37 vm07 bash[17031]: audit 2026-03-07T06:57:37.092281+0000 mon.vm07 (mon.0) 681 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:37.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:37 vm07 bash[17031]: audit 2026-03-07T06:57:37.093330+0000 mon.vm07 (mon.0) 682 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:37.596 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:37 vm07 bash[17031]: audit 2026-03-07T06:57:37.093330+0000 mon.vm07 (mon.0) 682 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:37.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:37 vm09 bash[21220]: cephadm 2026-03-07T06:57:36.195071+0000 mgr.vm07.yrfcuj (mgr.14201) 163 : cephadm [INF] Deploying daemon rgw.foo.vm09.fgzozy on vm09 2026-03-07T07:57:37.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:37 vm09 bash[21220]: cephadm 2026-03-07T06:57:36.195071+0000 mgr.vm07.yrfcuj (mgr.14201) 163 : cephadm [INF] Deploying daemon rgw.foo.vm09.fgzozy on vm09 2026-03-07T07:57:37.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:37 vm09 bash[21220]: audit 2026-03-07T06:57:37.071313+0000 mon.vm07 (mon.0) 676 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:37.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:37 vm09 bash[21220]: audit 2026-03-07T06:57:37.071313+0000 mon.vm07 (mon.0) 676 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:37.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:37 vm09 bash[21220]: audit 2026-03-07T06:57:37.075757+0000 mon.vm07 (mon.0) 677 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:37.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:37 vm09 bash[21220]: audit 2026-03-07T06:57:37.075757+0000 mon.vm07 (mon.0) 677 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:37.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:37 vm09 bash[21220]: audit 2026-03-07T06:57:37.079856+0000 mon.vm07 (mon.0) 678 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:37 vm09 bash[21220]: audit 2026-03-07T06:57:37.079856+0000 mon.vm07 (mon.0) 678 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:37 vm09 bash[21220]: audit 2026-03-07T06:57:37.080317+0000 mon.vm07 (mon.0) 679 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm07.bgfaka", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-07T07:57:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:37 vm09 bash[21220]: audit 2026-03-07T06:57:37.080317+0000 mon.vm07 (mon.0) 679 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm07.bgfaka", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-07T07:57:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:37 vm09 bash[21220]: audit 2026-03-07T06:57:37.087888+0000 mon.vm07 (mon.0) 680 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm07.bgfaka", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-07T07:57:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:37 vm09 bash[21220]: audit 2026-03-07T06:57:37.087888+0000 mon.vm07 (mon.0) 680 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm07.bgfaka", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-07T07:57:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:37 vm09 bash[21220]: audit 2026-03-07T06:57:37.092281+0000 mon.vm07 (mon.0) 681 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:37 vm09 bash[21220]: audit 2026-03-07T06:57:37.092281+0000 mon.vm07 (mon.0) 681 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:37 vm09 bash[21220]: audit 2026-03-07T06:57:37.093330+0000 mon.vm07 (mon.0) 682 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:37 vm09 bash[21220]: audit 2026-03-07T06:57:37.093330+0000 mon.vm07 (mon.0) 682 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:37.887 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:37 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:57:37.887 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:37 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:57:38.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: cephadm 2026-03-07T06:57:37.093814+0000 mgr.vm07.yrfcuj (mgr.14201) 164 : cephadm [INF] Deploying daemon rgw.foo.vm07.bgfaka on vm07 2026-03-07T07:57:38.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: cephadm 2026-03-07T06:57:37.093814+0000 mgr.vm07.yrfcuj (mgr.14201) 164 : cephadm [INF] Deploying daemon rgw.foo.vm07.bgfaka on vm07 2026-03-07T07:57:38.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: cluster 2026-03-07T06:57:37.109235+0000 mgr.vm07.yrfcuj (mgr.14201) 165 : cluster [DBG] pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:38.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: cluster 2026-03-07T06:57:37.109235+0000 mgr.vm07.yrfcuj (mgr.14201) 165 : cluster [DBG] pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:38.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:37.908197+0000 mon.vm07 (mon.0) 683 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:38.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:37.908197+0000 mon.vm07 (mon.0) 683 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:38.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:37.912628+0000 mon.vm07 (mon.0) 684 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:38.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:37.912628+0000 mon.vm07 (mon.0) 684 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:38.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:37.915769+0000 mon.vm07 (mon.0) 685 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:38.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:37.915769+0000 mon.vm07 (mon.0) 685 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:38.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:37.916228+0000 mon.vm07 (mon.0) 686 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm09.kpgoql", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-07T07:57:38.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:37.916228+0000 mon.vm07 (mon.0) 686 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm09.kpgoql", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-07T07:57:38.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:37.918416+0000 mon.vm07 (mon.0) 687 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm09.kpgoql", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-07T07:57:38.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:37.918416+0000 mon.vm07 (mon.0) 687 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm09.kpgoql", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-07T07:57:38.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:37.922493+0000 mon.vm07 (mon.0) 688 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:38.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:37.922493+0000 mon.vm07 (mon.0) 688 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:38.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:37.926419+0000 mon.vm07 (mon.0) 689 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:38.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:37.926419+0000 mon.vm07 (mon.0) 689 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:38.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: cluster 2026-03-07T06:57:38.100529+0000 mon.vm07 (mon.0) 690 : cluster [DBG] osdmap e24: 8 total, 8 up, 8 in 2026-03-07T07:57:38.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: cluster 2026-03-07T06:57:38.100529+0000 mon.vm07 (mon.0) 690 : cluster [DBG] osdmap e24: 8 total, 8 up, 8 in 2026-03-07T07:57:38.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:38.111496+0000 mon.vm07 (mon.0) 691 : audit [INF] from='client.? 192.168.123.107:0/548804138' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-07T07:57:38.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:38.111496+0000 mon.vm07 (mon.0) 691 : audit [INF] from='client.? 192.168.123.107:0/548804138' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-07T07:57:38.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:38.118678+0000 mon.vm07 (mon.0) 692 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-07T07:57:38.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:38.118678+0000 mon.vm07 (mon.0) 692 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-07T07:57:38.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:38.123047+0000 mon.vm09 (mon.1) 21 : audit [INF] from='client.? 192.168.123.109:0/1721027402' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-07T07:57:38.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 bash[21220]: audit 2026-03-07T06:57:38.123047+0000 mon.vm09 (mon.1) 21 : audit [INF] from='client.? 192.168.123.109:0/1721027402' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-07T07:57:38.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: cephadm 2026-03-07T06:57:37.093814+0000 mgr.vm07.yrfcuj (mgr.14201) 164 : cephadm [INF] Deploying daemon rgw.foo.vm07.bgfaka on vm07 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: cephadm 2026-03-07T06:57:37.093814+0000 mgr.vm07.yrfcuj (mgr.14201) 164 : cephadm [INF] Deploying daemon rgw.foo.vm07.bgfaka on vm07 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: cluster 2026-03-07T06:57:37.109235+0000 mgr.vm07.yrfcuj (mgr.14201) 165 : cluster [DBG] pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: cluster 2026-03-07T06:57:37.109235+0000 mgr.vm07.yrfcuj (mgr.14201) 165 : cluster [DBG] pgmap v107: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:37.908197+0000 mon.vm07 (mon.0) 683 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:37.908197+0000 mon.vm07 (mon.0) 683 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:37.912628+0000 mon.vm07 (mon.0) 684 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:37.912628+0000 mon.vm07 (mon.0) 684 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:37.915769+0000 mon.vm07 (mon.0) 685 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:37.915769+0000 mon.vm07 (mon.0) 685 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:37.916228+0000 mon.vm07 (mon.0) 686 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm09.kpgoql", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:37.916228+0000 mon.vm07 (mon.0) 686 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm09.kpgoql", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:37.918416+0000 mon.vm07 (mon.0) 687 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm09.kpgoql", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:37.918416+0000 mon.vm07 (mon.0) 687 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm09.kpgoql", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:37.922493+0000 mon.vm07 (mon.0) 688 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:37.922493+0000 mon.vm07 (mon.0) 688 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:37.926419+0000 mon.vm07 (mon.0) 689 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:37.926419+0000 mon.vm07 (mon.0) 689 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: cluster 2026-03-07T06:57:38.100529+0000 mon.vm07 (mon.0) 690 : cluster [DBG] osdmap e24: 8 total, 8 up, 8 in 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: cluster 2026-03-07T06:57:38.100529+0000 mon.vm07 (mon.0) 690 : cluster [DBG] osdmap e24: 8 total, 8 up, 8 in 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:38.111496+0000 mon.vm07 (mon.0) 691 : audit [INF] from='client.? 192.168.123.107:0/548804138' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:38.111496+0000 mon.vm07 (mon.0) 691 : audit [INF] from='client.? 192.168.123.107:0/548804138' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-07T07:57:38.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:38.118678+0000 mon.vm07 (mon.0) 692 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-07T07:57:38.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:38.118678+0000 mon.vm07 (mon.0) 692 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-07T07:57:38.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:38.123047+0000 mon.vm09 (mon.1) 21 : audit [INF] from='client.? 192.168.123.109:0/1721027402' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-07T07:57:38.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:38 vm07 bash[17031]: audit 2026-03-07T06:57:38.123047+0000 mon.vm09 (mon.1) 21 : audit [INF] from='client.? 192.168.123.109:0/1721027402' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-07T07:57:38.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:38 vm09 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: cephadm 2026-03-07T06:57:37.927021+0000 mgr.vm07.yrfcuj (mgr.14201) 166 : cephadm [INF] Deploying daemon rgw.foo.vm09.kpgoql on vm09 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: cephadm 2026-03-07T06:57:37.927021+0000 mgr.vm07.yrfcuj (mgr.14201) 166 : cephadm [INF] Deploying daemon rgw.foo.vm09.kpgoql on vm09 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: audit 2026-03-07T06:57:38.721619+0000 mon.vm07 (mon.0) 693 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: audit 2026-03-07T06:57:38.721619+0000 mon.vm07 (mon.0) 693 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: audit 2026-03-07T06:57:38.726002+0000 mon.vm07 (mon.0) 694 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: audit 2026-03-07T06:57:38.726002+0000 mon.vm07 (mon.0) 694 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: audit 2026-03-07T06:57:38.734647+0000 mon.vm07 (mon.0) 695 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: audit 2026-03-07T06:57:38.734647+0000 mon.vm07 (mon.0) 695 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: audit 2026-03-07T06:57:38.735040+0000 mon.vm07 (mon.0) 696 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm07.lrxyjy", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: audit 2026-03-07T06:57:38.735040+0000 mon.vm07 (mon.0) 696 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm07.lrxyjy", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: audit 2026-03-07T06:57:38.738957+0000 mon.vm07 (mon.0) 697 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm07.lrxyjy", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: audit 2026-03-07T06:57:38.738957+0000 mon.vm07 (mon.0) 697 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm07.lrxyjy", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: audit 2026-03-07T06:57:38.743264+0000 mon.vm07 (mon.0) 698 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: audit 2026-03-07T06:57:38.743264+0000 mon.vm07 (mon.0) 698 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: audit 2026-03-07T06:57:38.744213+0000 mon.vm07 (mon.0) 699 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: audit 2026-03-07T06:57:38.744213+0000 mon.vm07 (mon.0) 699 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: audit 2026-03-07T06:57:39.133238+0000 mon.vm07 (mon.0) 700 : audit [INF] from='client.? 192.168.123.107:0/548804138' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]': finished 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: audit 2026-03-07T06:57:39.133238+0000 mon.vm07 (mon.0) 700 : audit [INF] from='client.? 192.168.123.107:0/548804138' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]': finished 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: audit 2026-03-07T06:57:39.133270+0000 mon.vm07 (mon.0) 701 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]': finished 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: audit 2026-03-07T06:57:39.133270+0000 mon.vm07 (mon.0) 701 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]': finished 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: cluster 2026-03-07T06:57:39.144483+0000 mon.vm07 (mon.0) 702 : cluster [DBG] osdmap e25: 8 total, 8 up, 8 in 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 bash[17031]: cluster 2026-03-07T06:57:39.144483+0000 mon.vm07 (mon.0) 702 : cluster [DBG] osdmap e25: 8 total, 8 up, 8 in 2026-03-07T07:57:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: cephadm 2026-03-07T06:57:37.927021+0000 mgr.vm07.yrfcuj (mgr.14201) 166 : cephadm [INF] Deploying daemon rgw.foo.vm09.kpgoql on vm09 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: cephadm 2026-03-07T06:57:37.927021+0000 mgr.vm07.yrfcuj (mgr.14201) 166 : cephadm [INF] Deploying daemon rgw.foo.vm09.kpgoql on vm09 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: audit 2026-03-07T06:57:38.721619+0000 mon.vm07 (mon.0) 693 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: audit 2026-03-07T06:57:38.721619+0000 mon.vm07 (mon.0) 693 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: audit 2026-03-07T06:57:38.726002+0000 mon.vm07 (mon.0) 694 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: audit 2026-03-07T06:57:38.726002+0000 mon.vm07 (mon.0) 694 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: audit 2026-03-07T06:57:38.734647+0000 mon.vm07 (mon.0) 695 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: audit 2026-03-07T06:57:38.734647+0000 mon.vm07 (mon.0) 695 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: audit 2026-03-07T06:57:38.735040+0000 mon.vm07 (mon.0) 696 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm07.lrxyjy", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: audit 2026-03-07T06:57:38.735040+0000 mon.vm07 (mon.0) 696 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm07.lrxyjy", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: audit 2026-03-07T06:57:38.738957+0000 mon.vm07 (mon.0) 697 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm07.lrxyjy", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: audit 2026-03-07T06:57:38.738957+0000 mon.vm07 (mon.0) 697 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foo.vm07.lrxyjy", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: audit 2026-03-07T06:57:38.743264+0000 mon.vm07 (mon.0) 698 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: audit 2026-03-07T06:57:38.743264+0000 mon.vm07 (mon.0) 698 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: audit 2026-03-07T06:57:38.744213+0000 mon.vm07 (mon.0) 699 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: audit 2026-03-07T06:57:38.744213+0000 mon.vm07 (mon.0) 699 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: audit 2026-03-07T06:57:39.133238+0000 mon.vm07 (mon.0) 700 : audit [INF] from='client.? 192.168.123.107:0/548804138' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]': finished 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: audit 2026-03-07T06:57:39.133238+0000 mon.vm07 (mon.0) 700 : audit [INF] from='client.? 192.168.123.107:0/548804138' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]': finished 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: audit 2026-03-07T06:57:39.133270+0000 mon.vm07 (mon.0) 701 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]': finished 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: audit 2026-03-07T06:57:39.133270+0000 mon.vm07 (mon.0) 701 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]': finished 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: cluster 2026-03-07T06:57:39.144483+0000 mon.vm07 (mon.0) 702 : cluster [DBG] osdmap e25: 8 total, 8 up, 8 in 2026-03-07T07:57:39.869 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:39 vm09 bash[21220]: cluster 2026-03-07T06:57:39.144483+0000 mon.vm07 (mon.0) 702 : cluster [DBG] osdmap e25: 8 total, 8 up, 8 in 2026-03-07T07:57:39.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:39 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: cephadm 2026-03-07T06:57:38.744647+0000 mgr.vm07.yrfcuj (mgr.14201) 167 : cephadm [INF] Deploying daemon rgw.foo.vm07.lrxyjy on vm07 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: cephadm 2026-03-07T06:57:38.744647+0000 mgr.vm07.yrfcuj (mgr.14201) 167 : cephadm [INF] Deploying daemon rgw.foo.vm07.lrxyjy on vm07 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: cluster 2026-03-07T06:57:39.109533+0000 mgr.vm07.yrfcuj (mgr.14201) 168 : cluster [DBG] pgmap v109: 33 pgs: 5 creating+peering, 27 unknown, 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: cluster 2026-03-07T06:57:39.109533+0000 mgr.vm07.yrfcuj (mgr.14201) 168 : cluster [DBG] pgmap v109: 33 pgs: 5 creating+peering, 27 unknown, 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: cluster 2026-03-07T06:57:39.742732+0000 mon.vm07 (mon.0) 703 : cluster [WRN] Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: cluster 2026-03-07T06:57:39.742732+0000 mon.vm07 (mon.0) 703 : cluster [WRN] Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:39.759564+0000 mon.vm07 (mon.0) 704 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:39.759564+0000 mon.vm07 (mon.0) 704 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:39.764890+0000 mon.vm07 (mon.0) 705 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:39.764890+0000 mon.vm07 (mon.0) 705 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:39.774277+0000 mon.vm07 (mon.0) 706 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:39.774277+0000 mon.vm07 (mon.0) 706 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:39.777327+0000 mon.vm07 (mon.0) 707 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:39.777327+0000 mon.vm07 (mon.0) 707 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:39.780246+0000 mon.vm07 (mon.0) 708 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:39.780246+0000 mon.vm07 (mon.0) 708 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:39.784496+0000 mon.vm07 (mon.0) 709 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:39.784496+0000 mon.vm07 (mon.0) 709 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:40.130932+0000 mon.vm07 (mon.0) 710 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:40.130932+0000 mon.vm07 (mon.0) 710 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:40.131488+0000 mon.vm07 (mon.0) 711 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:40.131488+0000 mon.vm07 (mon.0) 711 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: cluster 2026-03-07T06:57:40.140717+0000 mon.vm07 (mon.0) 712 : cluster [DBG] osdmap e26: 8 total, 8 up, 8 in 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: cluster 2026-03-07T06:57:40.140717+0000 mon.vm07 (mon.0) 712 : cluster [DBG] osdmap e26: 8 total, 8 up, 8 in 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:40.142360+0000 mon.vm07 (mon.0) 713 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.385 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:40.142360+0000 mon.vm07 (mon.0) 713 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.386 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:40.142505+0000 mon.vm07 (mon.0) 714 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.386 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:40.142505+0000 mon.vm07 (mon.0) 714 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.386 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:40.195686+0000 mon.vm07 (mon.0) 715 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.386 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:40.195686+0000 mon.vm07 (mon.0) 715 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.386 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:40.196460+0000 mon.vm07 (mon.0) 716 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.386 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:40.196460+0000 mon.vm07 (mon.0) 716 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.386 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:40.200491+0000 mon.vm09 (mon.1) 22 : audit [INF] from='client.? 192.168.123.109:0/357189846' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.386 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:40.200491+0000 mon.vm09 (mon.1) 22 : audit [INF] from='client.? 192.168.123.109:0/357189846' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.386 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:40.201327+0000 mon.vm09 (mon.1) 23 : audit [INF] from='client.? 192.168.123.109:0/1001165862' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.386 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:40 vm09 bash[21220]: audit 2026-03-07T06:57:40.201327+0000 mon.vm09 (mon.1) 23 : audit [INF] from='client.? 192.168.123.109:0/1001165862' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: cephadm 2026-03-07T06:57:38.744647+0000 mgr.vm07.yrfcuj (mgr.14201) 167 : cephadm [INF] Deploying daemon rgw.foo.vm07.lrxyjy on vm07 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: cephadm 2026-03-07T06:57:38.744647+0000 mgr.vm07.yrfcuj (mgr.14201) 167 : cephadm [INF] Deploying daemon rgw.foo.vm07.lrxyjy on vm07 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: cluster 2026-03-07T06:57:39.109533+0000 mgr.vm07.yrfcuj (mgr.14201) 168 : cluster [DBG] pgmap v109: 33 pgs: 5 creating+peering, 27 unknown, 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: cluster 2026-03-07T06:57:39.109533+0000 mgr.vm07.yrfcuj (mgr.14201) 168 : cluster [DBG] pgmap v109: 33 pgs: 5 creating+peering, 27 unknown, 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: cluster 2026-03-07T06:57:39.742732+0000 mon.vm07 (mon.0) 703 : cluster [WRN] Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: cluster 2026-03-07T06:57:39.742732+0000 mon.vm07 (mon.0) 703 : cluster [WRN] Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:39.759564+0000 mon.vm07 (mon.0) 704 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:39.759564+0000 mon.vm07 (mon.0) 704 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:39.764890+0000 mon.vm07 (mon.0) 705 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:39.764890+0000 mon.vm07 (mon.0) 705 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:39.774277+0000 mon.vm07 (mon.0) 706 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:39.774277+0000 mon.vm07 (mon.0) 706 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:39.777327+0000 mon.vm07 (mon.0) 707 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:39.777327+0000 mon.vm07 (mon.0) 707 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:39.780246+0000 mon.vm07 (mon.0) 708 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:39.780246+0000 mon.vm07 (mon.0) 708 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:39.784496+0000 mon.vm07 (mon.0) 709 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:39.784496+0000 mon.vm07 (mon.0) 709 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:40.130932+0000 mon.vm07 (mon.0) 710 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:40.130932+0000 mon.vm07 (mon.0) 710 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:40.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:40.131488+0000 mon.vm07 (mon.0) 711 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:57:40.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:40.131488+0000 mon.vm07 (mon.0) 711 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:57:40.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: cluster 2026-03-07T06:57:40.140717+0000 mon.vm07 (mon.0) 712 : cluster [DBG] osdmap e26: 8 total, 8 up, 8 in 2026-03-07T07:57:40.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: cluster 2026-03-07T06:57:40.140717+0000 mon.vm07 (mon.0) 712 : cluster [DBG] osdmap e26: 8 total, 8 up, 8 in 2026-03-07T07:57:40.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:40.142360+0000 mon.vm07 (mon.0) 713 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:40.142360+0000 mon.vm07 (mon.0) 713 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:40.142505+0000 mon.vm07 (mon.0) 714 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:40.142505+0000 mon.vm07 (mon.0) 714 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:40.195686+0000 mon.vm07 (mon.0) 715 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:40.195686+0000 mon.vm07 (mon.0) 715 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:40.196460+0000 mon.vm07 (mon.0) 716 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:40.196460+0000 mon.vm07 (mon.0) 716 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:40.200491+0000 mon.vm09 (mon.1) 22 : audit [INF] from='client.? 192.168.123.109:0/357189846' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:40.200491+0000 mon.vm09 (mon.1) 22 : audit [INF] from='client.? 192.168.123.109:0/357189846' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:40.201327+0000 mon.vm09 (mon.1) 23 : audit [INF] from='client.? 192.168.123.109:0/1001165862' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:40.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:40 vm07 bash[17031]: audit 2026-03-07T06:57:40.201327+0000 mon.vm09 (mon.1) 23 : audit [INF] from='client.? 192.168.123.109:0/1001165862' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-07T07:57:41.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:41 vm07 bash[17031]: cephadm 2026-03-07T06:57:39.774714+0000 mgr.vm07.yrfcuj (mgr.14201) 169 : cephadm [INF] Saving service rgw.foo spec with placement count:4;* 2026-03-07T07:57:41.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:41 vm07 bash[17031]: cephadm 2026-03-07T06:57:39.774714+0000 mgr.vm07.yrfcuj (mgr.14201) 169 : cephadm [INF] Saving service rgw.foo spec with placement count:4;* 2026-03-07T07:57:41.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:41 vm07 bash[17031]: cephadm 2026-03-07T06:57:39.789493+0000 mgr.vm07.yrfcuj (mgr.14201) 170 : cephadm [INF] Deploying daemon haproxy.rgw.foo.vm09.hhbnqk on vm09 2026-03-07T07:57:41.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:41 vm07 bash[17031]: cephadm 2026-03-07T06:57:39.789493+0000 mgr.vm07.yrfcuj (mgr.14201) 170 : cephadm [INF] Deploying daemon haproxy.rgw.foo.vm09.hhbnqk on vm09 2026-03-07T07:57:41.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:41 vm07 bash[17031]: audit 2026-03-07T06:57:41.141872+0000 mon.vm07 (mon.0) 717 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-07T07:57:41.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:41 vm07 bash[17031]: audit 2026-03-07T06:57:41.141872+0000 mon.vm07 (mon.0) 717 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-07T07:57:41.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:41 vm07 bash[17031]: audit 2026-03-07T06:57:41.141928+0000 mon.vm07 (mon.0) 718 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-07T07:57:41.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:41 vm07 bash[17031]: audit 2026-03-07T06:57:41.141928+0000 mon.vm07 (mon.0) 718 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-07T07:57:41.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:41 vm07 bash[17031]: audit 2026-03-07T06:57:41.142014+0000 mon.vm07 (mon.0) 719 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-07T07:57:41.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:41 vm07 bash[17031]: audit 2026-03-07T06:57:41.142014+0000 mon.vm07 (mon.0) 719 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-07T07:57:41.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:41 vm07 bash[17031]: audit 2026-03-07T06:57:41.142039+0000 mon.vm07 (mon.0) 720 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-07T07:57:41.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:41 vm07 bash[17031]: audit 2026-03-07T06:57:41.142039+0000 mon.vm07 (mon.0) 720 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-07T07:57:41.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:41 vm07 bash[17031]: cluster 2026-03-07T06:57:41.145809+0000 mon.vm07 (mon.0) 721 : cluster [DBG] osdmap e27: 8 total, 8 up, 8 in 2026-03-07T07:57:41.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:41 vm07 bash[17031]: cluster 2026-03-07T06:57:41.145809+0000 mon.vm07 (mon.0) 721 : cluster [DBG] osdmap e27: 8 total, 8 up, 8 in 2026-03-07T07:57:41.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:41 vm09 bash[21220]: cephadm 2026-03-07T06:57:39.774714+0000 mgr.vm07.yrfcuj (mgr.14201) 169 : cephadm [INF] Saving service rgw.foo spec with placement count:4;* 2026-03-07T07:57:41.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:41 vm09 bash[21220]: cephadm 2026-03-07T06:57:39.774714+0000 mgr.vm07.yrfcuj (mgr.14201) 169 : cephadm [INF] Saving service rgw.foo spec with placement count:4;* 2026-03-07T07:57:41.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:41 vm09 bash[21220]: cephadm 2026-03-07T06:57:39.789493+0000 mgr.vm07.yrfcuj (mgr.14201) 170 : cephadm [INF] Deploying daemon haproxy.rgw.foo.vm09.hhbnqk on vm09 2026-03-07T07:57:41.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:41 vm09 bash[21220]: cephadm 2026-03-07T06:57:39.789493+0000 mgr.vm07.yrfcuj (mgr.14201) 170 : cephadm [INF] Deploying daemon haproxy.rgw.foo.vm09.hhbnqk on vm09 2026-03-07T07:57:41.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:41 vm09 bash[21220]: audit 2026-03-07T06:57:41.141872+0000 mon.vm07 (mon.0) 717 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-07T07:57:41.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:41 vm09 bash[21220]: audit 2026-03-07T06:57:41.141872+0000 mon.vm07 (mon.0) 717 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-07T07:57:41.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:41 vm09 bash[21220]: audit 2026-03-07T06:57:41.141928+0000 mon.vm07 (mon.0) 718 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-07T07:57:41.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:41 vm09 bash[21220]: audit 2026-03-07T06:57:41.141928+0000 mon.vm07 (mon.0) 718 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-07T07:57:41.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:41 vm09 bash[21220]: audit 2026-03-07T06:57:41.142014+0000 mon.vm07 (mon.0) 719 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-07T07:57:41.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:41 vm09 bash[21220]: audit 2026-03-07T06:57:41.142014+0000 mon.vm07 (mon.0) 719 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-07T07:57:41.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:41 vm09 bash[21220]: audit 2026-03-07T06:57:41.142039+0000 mon.vm07 (mon.0) 720 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-07T07:57:41.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:41 vm09 bash[21220]: audit 2026-03-07T06:57:41.142039+0000 mon.vm07 (mon.0) 720 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-07T07:57:41.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:41 vm09 bash[21220]: cluster 2026-03-07T06:57:41.145809+0000 mon.vm07 (mon.0) 721 : cluster [DBG] osdmap e27: 8 total, 8 up, 8 in 2026-03-07T07:57:41.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:41 vm09 bash[21220]: cluster 2026-03-07T06:57:41.145809+0000 mon.vm07 (mon.0) 721 : cluster [DBG] osdmap e27: 8 total, 8 up, 8 in 2026-03-07T07:57:42.079 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:57:42.470 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:57:42.470 INFO:teuthology.orchestra.run.vm07.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-07T06:53:25.695727Z", "last_refresh": "2026-03-07T06:57:30.067219Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-07T06:54:32.927755Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-07T06:53:24.247577Z", "last_refresh": "2026-03-07T06:57:29.595811Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:33.727624Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-07T06:53:23.837568Z", "last_refresh": "2026-03-07T06:57:29.595876Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-07T06:53:24.987960Z", "last_refresh": "2026-03-07T06:57:30.067033Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-07T06:57:29.850523Z service:ingress.rgw.foo [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "rgw.foo", "service_name": "ingress.rgw.foo", "service_type": "ingress", "spec": {"backend_service": "rgw.foo", "first_virtual_router_id": 50, "frontend_port": 9000, "monitor_port": 9001, "virtual_ip": "12.12.1.107/22"}, "status": {"created": "2026-03-07T06:57:29.846008Z", "ports": [9000, 9001], "running": 0, "size": 4, "virtual_ip": "12.12.1.107/22"}}, {"events": ["2026-03-07T06:54:35.185395Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-07T06:53:23.396871Z", "last_refresh": "2026-03-07T06:57:29.595909Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:36.226821Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm07:192.168.123.107=vm07", "vm09:192.168.123.109=vm09"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-07T06:54:13.493474Z", "last_refresh": "2026-03-07T06:57:29.595844Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:34.440095Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-07T06:53:25.333423Z", "last_refresh": "2026-03-07T06:57:29.596000Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:54.121213Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-07T06:54:54.116553Z", "last_refresh": "2026-03-07T06:57:29.595721Z", "running": 8, "size": 8}}, {"events": ["2026-03-07T06:54:36.229563Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-07T06:53:24.610265Z", "last_refresh": "2026-03-07T06:57:30.066979Z", "ports": [9095], "running": 1, "size": 1}}, {"events": ["2026-03-07T06:57:39.780378Z service:rgw.foo [INFO] \"service was created\""], "placement": {"count": 4, "host_pattern": "*"}, "service_id": "foo", "service_name": "rgw.foo", "service_type": "rgw", "spec": {"rgw_frontend_port": 8000}, "status": {"created": "2026-03-07T06:57:39.774719Z", "ports": [8000], "running": 0, "size": 4}}] 2026-03-07T07:57:42.532 INFO:tasks.cephadm:rgw.foo has 0/4 2026-03-07T07:57:42.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:42 vm07 bash[17031]: cluster 2026-03-07T06:57:41.109847+0000 mgr.vm07.yrfcuj (mgr.14201) 171 : cluster [DBG] pgmap v112: 65 pgs: 5 creating+peering, 53 unknown, 7 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 1.5 KiB/s rd, 170 B/s wr, 1 op/s 2026-03-07T07:57:42.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:42 vm07 bash[17031]: cluster 2026-03-07T06:57:41.109847+0000 mgr.vm07.yrfcuj (mgr.14201) 171 : cluster [DBG] pgmap v112: 65 pgs: 5 creating+peering, 53 unknown, 7 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 1.5 KiB/s rd, 170 B/s wr, 1 op/s 2026-03-07T07:57:42.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:42 vm07 bash[17031]: audit 2026-03-07T06:57:41.637651+0000 mon.vm07 (mon.0) 722 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:42.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:42 vm07 bash[17031]: audit 2026-03-07T06:57:41.637651+0000 mon.vm07 (mon.0) 722 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:42.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:42 vm07 bash[17031]: cluster 2026-03-07T06:57:42.147549+0000 mon.vm07 (mon.0) 723 : cluster [DBG] osdmap e28: 8 total, 8 up, 8 in 2026-03-07T07:57:42.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:42 vm07 bash[17031]: cluster 2026-03-07T06:57:42.147549+0000 mon.vm07 (mon.0) 723 : cluster [DBG] osdmap e28: 8 total, 8 up, 8 in 2026-03-07T07:57:42.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:42 vm07 bash[17031]: audit 2026-03-07T06:57:42.149460+0000 mon.vm07 (mon.0) 724 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:42 vm07 bash[17031]: audit 2026-03-07T06:57:42.149460+0000 mon.vm07 (mon.0) 724 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:42 vm07 bash[17031]: audit 2026-03-07T06:57:42.149519+0000 mon.vm07 (mon.0) 725 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:42 vm07 bash[17031]: audit 2026-03-07T06:57:42.149519+0000 mon.vm07 (mon.0) 725 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:42 vm07 bash[17031]: audit 2026-03-07T06:57:42.164341+0000 mon.vm07 (mon.0) 726 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:42 vm07 bash[17031]: audit 2026-03-07T06:57:42.164341+0000 mon.vm07 (mon.0) 726 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:42 vm07 bash[17031]: audit 2026-03-07T06:57:42.164907+0000 mon.vm07 (mon.0) 727 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:42 vm07 bash[17031]: audit 2026-03-07T06:57:42.164907+0000 mon.vm07 (mon.0) 727 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:42 vm07 bash[17031]: audit 2026-03-07T06:57:42.169158+0000 mon.vm09 (mon.1) 24 : audit [INF] from='client.? 192.168.123.109:0/357189846' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:42 vm07 bash[17031]: audit 2026-03-07T06:57:42.169158+0000 mon.vm09 (mon.1) 24 : audit [INF] from='client.? 192.168.123.109:0/357189846' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:42 vm07 bash[17031]: audit 2026-03-07T06:57:42.169767+0000 mon.vm09 (mon.1) 25 : audit [INF] from='client.? 192.168.123.109:0/1001165862' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:42 vm07 bash[17031]: audit 2026-03-07T06:57:42.169767+0000 mon.vm09 (mon.1) 25 : audit [INF] from='client.? 192.168.123.109:0/1001165862' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:42 vm09 bash[21220]: cluster 2026-03-07T06:57:41.109847+0000 mgr.vm07.yrfcuj (mgr.14201) 171 : cluster [DBG] pgmap v112: 65 pgs: 5 creating+peering, 53 unknown, 7 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 1.5 KiB/s rd, 170 B/s wr, 1 op/s 2026-03-07T07:57:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:42 vm09 bash[21220]: cluster 2026-03-07T06:57:41.109847+0000 mgr.vm07.yrfcuj (mgr.14201) 171 : cluster [DBG] pgmap v112: 65 pgs: 5 creating+peering, 53 unknown, 7 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 1.5 KiB/s rd, 170 B/s wr, 1 op/s 2026-03-07T07:57:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:42 vm09 bash[21220]: audit 2026-03-07T06:57:41.637651+0000 mon.vm07 (mon.0) 722 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:42 vm09 bash[21220]: audit 2026-03-07T06:57:41.637651+0000 mon.vm07 (mon.0) 722 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:42 vm09 bash[21220]: cluster 2026-03-07T06:57:42.147549+0000 mon.vm07 (mon.0) 723 : cluster [DBG] osdmap e28: 8 total, 8 up, 8 in 2026-03-07T07:57:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:42 vm09 bash[21220]: cluster 2026-03-07T06:57:42.147549+0000 mon.vm07 (mon.0) 723 : cluster [DBG] osdmap e28: 8 total, 8 up, 8 in 2026-03-07T07:57:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:42 vm09 bash[21220]: audit 2026-03-07T06:57:42.149460+0000 mon.vm07 (mon.0) 724 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:42 vm09 bash[21220]: audit 2026-03-07T06:57:42.149460+0000 mon.vm07 (mon.0) 724 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:42 vm09 bash[21220]: audit 2026-03-07T06:57:42.149519+0000 mon.vm07 (mon.0) 725 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:42 vm09 bash[21220]: audit 2026-03-07T06:57:42.149519+0000 mon.vm07 (mon.0) 725 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:42 vm09 bash[21220]: audit 2026-03-07T06:57:42.164341+0000 mon.vm07 (mon.0) 726 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:42 vm09 bash[21220]: audit 2026-03-07T06:57:42.164341+0000 mon.vm07 (mon.0) 726 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:42 vm09 bash[21220]: audit 2026-03-07T06:57:42.164907+0000 mon.vm07 (mon.0) 727 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:42 vm09 bash[21220]: audit 2026-03-07T06:57:42.164907+0000 mon.vm07 (mon.0) 727 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:42 vm09 bash[21220]: audit 2026-03-07T06:57:42.169158+0000 mon.vm09 (mon.1) 24 : audit [INF] from='client.? 192.168.123.109:0/357189846' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:42 vm09 bash[21220]: audit 2026-03-07T06:57:42.169158+0000 mon.vm09 (mon.1) 24 : audit [INF] from='client.? 192.168.123.109:0/357189846' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:42 vm09 bash[21220]: audit 2026-03-07T06:57:42.169767+0000 mon.vm09 (mon.1) 25 : audit [INF] from='client.? 192.168.123.109:0/1001165862' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:42 vm09 bash[21220]: audit 2026-03-07T06:57:42.169767+0000 mon.vm09 (mon.1) 25 : audit [INF] from='client.? 192.168.123.109:0/1001165862' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-07T07:57:43.532 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph orch ls -f json 2026-03-07T07:57:43.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:43 vm07 bash[17031]: audit 2026-03-07T06:57:42.462385+0000 mgr.vm07.yrfcuj (mgr.14201) 172 : audit [DBG] from='client.14538 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:57:43.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:43 vm07 bash[17031]: audit 2026-03-07T06:57:42.462385+0000 mgr.vm07.yrfcuj (mgr.14201) 172 : audit [DBG] from='client.14538 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:57:43.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:43 vm07 bash[17031]: audit 2026-03-07T06:57:43.151053+0000 mon.vm07 (mon.0) 728 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-07T07:57:43.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:43 vm07 bash[17031]: audit 2026-03-07T06:57:43.151053+0000 mon.vm07 (mon.0) 728 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-07T07:57:43.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:43 vm07 bash[17031]: audit 2026-03-07T06:57:43.151114+0000 mon.vm07 (mon.0) 729 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-07T07:57:43.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:43 vm07 bash[17031]: audit 2026-03-07T06:57:43.151114+0000 mon.vm07 (mon.0) 729 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-07T07:57:43.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:43 vm07 bash[17031]: audit 2026-03-07T06:57:43.151214+0000 mon.vm07 (mon.0) 730 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-07T07:57:43.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:43 vm07 bash[17031]: audit 2026-03-07T06:57:43.151214+0000 mon.vm07 (mon.0) 730 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-07T07:57:43.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:43 vm07 bash[17031]: audit 2026-03-07T06:57:43.151261+0000 mon.vm07 (mon.0) 731 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-07T07:57:43.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:43 vm07 bash[17031]: audit 2026-03-07T06:57:43.151261+0000 mon.vm07 (mon.0) 731 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-07T07:57:43.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:43 vm07 bash[17031]: cluster 2026-03-07T06:57:43.153442+0000 mon.vm07 (mon.0) 732 : cluster [DBG] osdmap e29: 8 total, 8 up, 8 in 2026-03-07T07:57:43.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:43 vm07 bash[17031]: cluster 2026-03-07T06:57:43.153442+0000 mon.vm07 (mon.0) 732 : cluster [DBG] osdmap e29: 8 total, 8 up, 8 in 2026-03-07T07:57:43.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:43 vm09 bash[21220]: audit 2026-03-07T06:57:42.462385+0000 mgr.vm07.yrfcuj (mgr.14201) 172 : audit [DBG] from='client.14538 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:57:43.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:43 vm09 bash[21220]: audit 2026-03-07T06:57:42.462385+0000 mgr.vm07.yrfcuj (mgr.14201) 172 : audit [DBG] from='client.14538 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:57:43.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:43 vm09 bash[21220]: audit 2026-03-07T06:57:43.151053+0000 mon.vm07 (mon.0) 728 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-07T07:57:43.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:43 vm09 bash[21220]: audit 2026-03-07T06:57:43.151053+0000 mon.vm07 (mon.0) 728 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-07T07:57:43.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:43 vm09 bash[21220]: audit 2026-03-07T06:57:43.151114+0000 mon.vm07 (mon.0) 729 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-07T07:57:43.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:43 vm09 bash[21220]: audit 2026-03-07T06:57:43.151114+0000 mon.vm07 (mon.0) 729 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-07T07:57:43.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:43 vm09 bash[21220]: audit 2026-03-07T06:57:43.151214+0000 mon.vm07 (mon.0) 730 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-07T07:57:43.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:43 vm09 bash[21220]: audit 2026-03-07T06:57:43.151214+0000 mon.vm07 (mon.0) 730 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-07T07:57:43.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:43 vm09 bash[21220]: audit 2026-03-07T06:57:43.151261+0000 mon.vm07 (mon.0) 731 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-07T07:57:43.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:43 vm09 bash[21220]: audit 2026-03-07T06:57:43.151261+0000 mon.vm07 (mon.0) 731 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-07T07:57:43.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:43 vm09 bash[21220]: cluster 2026-03-07T06:57:43.153442+0000 mon.vm07 (mon.0) 732 : cluster [DBG] osdmap e29: 8 total, 8 up, 8 in 2026-03-07T07:57:43.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:43 vm09 bash[21220]: cluster 2026-03-07T06:57:43.153442+0000 mon.vm07 (mon.0) 732 : cluster [DBG] osdmap e29: 8 total, 8 up, 8 in 2026-03-07T07:57:44.641 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:44 vm09 bash[21220]: cluster 2026-03-07T06:57:43.110316+0000 mgr.vm07.yrfcuj (mgr.14201) 173 : cluster [DBG] pgmap v115: 97 pgs: 23 creating+peering, 20 unknown, 54 active+clean; 450 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 5.7 KiB/s rd, 1.5 KiB/s wr, 9 op/s 2026-03-07T07:57:44.641 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:44 vm09 bash[21220]: cluster 2026-03-07T06:57:43.110316+0000 mgr.vm07.yrfcuj (mgr.14201) 173 : cluster [DBG] pgmap v115: 97 pgs: 23 creating+peering, 20 unknown, 54 active+clean; 450 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 5.7 KiB/s rd, 1.5 KiB/s wr, 9 op/s 2026-03-07T07:57:44.641 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:44 vm09 bash[21220]: cluster 2026-03-07T06:57:44.206756+0000 mon.vm07 (mon.0) 733 : cluster [DBG] osdmap e30: 8 total, 8 up, 8 in 2026-03-07T07:57:44.641 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:44 vm09 bash[21220]: cluster 2026-03-07T06:57:44.206756+0000 mon.vm07 (mon.0) 733 : cluster [DBG] osdmap e30: 8 total, 8 up, 8 in 2026-03-07T07:57:44.641 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:44 vm09 bash[21220]: audit 2026-03-07T06:57:44.208480+0000 mon.vm07 (mon.0) 734 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.641 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:44 vm09 bash[21220]: audit 2026-03-07T06:57:44.208480+0000 mon.vm07 (mon.0) 734 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.641 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:44 vm09 bash[21220]: audit 2026-03-07T06:57:44.208631+0000 mon.vm07 (mon.0) 735 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.641 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:44 vm09 bash[21220]: audit 2026-03-07T06:57:44.208631+0000 mon.vm07 (mon.0) 735 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.641 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:44 vm09 bash[21220]: audit 2026-03-07T06:57:44.268754+0000 mon.vm07 (mon.0) 736 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.641 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:44 vm09 bash[21220]: audit 2026-03-07T06:57:44.268754+0000 mon.vm07 (mon.0) 736 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.641 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:44 vm09 bash[21220]: audit 2026-03-07T06:57:44.269987+0000 mon.vm07 (mon.0) 737 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.641 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:44 vm09 bash[21220]: audit 2026-03-07T06:57:44.269987+0000 mon.vm07 (mon.0) 737 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.641 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:44 vm09 bash[21220]: audit 2026-03-07T06:57:44.273509+0000 mon.vm09 (mon.1) 26 : audit [INF] from='client.? 192.168.123.109:0/357189846' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.641 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:44 vm09 bash[21220]: audit 2026-03-07T06:57:44.273509+0000 mon.vm09 (mon.1) 26 : audit [INF] from='client.? 192.168.123.109:0/357189846' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.641 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:44 vm09 bash[21220]: audit 2026-03-07T06:57:44.273975+0000 mon.vm09 (mon.1) 27 : audit [INF] from='client.? 192.168.123.109:0/1001165862' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.641 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:44 vm09 bash[21220]: audit 2026-03-07T06:57:44.273975+0000 mon.vm09 (mon.1) 27 : audit [INF] from='client.? 192.168.123.109:0/1001165862' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:44 vm07 bash[17031]: cluster 2026-03-07T06:57:43.110316+0000 mgr.vm07.yrfcuj (mgr.14201) 173 : cluster [DBG] pgmap v115: 97 pgs: 23 creating+peering, 20 unknown, 54 active+clean; 450 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 5.7 KiB/s rd, 1.5 KiB/s wr, 9 op/s 2026-03-07T07:57:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:44 vm07 bash[17031]: cluster 2026-03-07T06:57:43.110316+0000 mgr.vm07.yrfcuj (mgr.14201) 173 : cluster [DBG] pgmap v115: 97 pgs: 23 creating+peering, 20 unknown, 54 active+clean; 450 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 5.7 KiB/s rd, 1.5 KiB/s wr, 9 op/s 2026-03-07T07:57:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:44 vm07 bash[17031]: cluster 2026-03-07T06:57:44.206756+0000 mon.vm07 (mon.0) 733 : cluster [DBG] osdmap e30: 8 total, 8 up, 8 in 2026-03-07T07:57:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:44 vm07 bash[17031]: cluster 2026-03-07T06:57:44.206756+0000 mon.vm07 (mon.0) 733 : cluster [DBG] osdmap e30: 8 total, 8 up, 8 in 2026-03-07T07:57:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:44 vm07 bash[17031]: audit 2026-03-07T06:57:44.208480+0000 mon.vm07 (mon.0) 734 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:44 vm07 bash[17031]: audit 2026-03-07T06:57:44.208480+0000 mon.vm07 (mon.0) 734 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:44 vm07 bash[17031]: audit 2026-03-07T06:57:44.208631+0000 mon.vm07 (mon.0) 735 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:44 vm07 bash[17031]: audit 2026-03-07T06:57:44.208631+0000 mon.vm07 (mon.0) 735 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:44 vm07 bash[17031]: audit 2026-03-07T06:57:44.268754+0000 mon.vm07 (mon.0) 736 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:44 vm07 bash[17031]: audit 2026-03-07T06:57:44.268754+0000 mon.vm07 (mon.0) 736 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:44 vm07 bash[17031]: audit 2026-03-07T06:57:44.269987+0000 mon.vm07 (mon.0) 737 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:44 vm07 bash[17031]: audit 2026-03-07T06:57:44.269987+0000 mon.vm07 (mon.0) 737 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:44 vm07 bash[17031]: audit 2026-03-07T06:57:44.273509+0000 mon.vm09 (mon.1) 26 : audit [INF] from='client.? 192.168.123.109:0/357189846' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:44 vm07 bash[17031]: audit 2026-03-07T06:57:44.273509+0000 mon.vm09 (mon.1) 26 : audit [INF] from='client.? 192.168.123.109:0/357189846' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:44 vm07 bash[17031]: audit 2026-03-07T06:57:44.273975+0000 mon.vm09 (mon.1) 27 : audit [INF] from='client.? 192.168.123.109:0/1001165862' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:44 vm07 bash[17031]: audit 2026-03-07T06:57:44.273975+0000 mon.vm09 (mon.1) 27 : audit [INF] from='client.? 192.168.123.109:0/1001165862' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-07T07:57:44.945 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:44 vm09 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:57:44.945 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:44 vm09 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:44.968380+0000 mon.vm07 (mon.0) 738 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:44.968380+0000 mon.vm07 (mon.0) 738 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:44.973840+0000 mon.vm07 (mon.0) 739 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:44.973840+0000 mon.vm07 (mon.0) 739 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:44.978192+0000 mon.vm07 (mon.0) 740 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:44.978192+0000 mon.vm07 (mon.0) 740 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: cephadm 2026-03-07T06:57:44.979654+0000 mgr.vm07.yrfcuj (mgr.14201) 174 : cephadm [INF] Deploying daemon haproxy.rgw.foo.vm07.rzyepz on vm07 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: cephadm 2026-03-07T06:57:44.979654+0000 mgr.vm07.yrfcuj (mgr.14201) 174 : cephadm [INF] Deploying daemon haproxy.rgw.foo.vm07.rzyepz on vm07 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: cluster 2026-03-07T06:57:45.110683+0000 mgr.vm07.yrfcuj (mgr.14201) 175 : cluster [DBG] pgmap v118: 129 pgs: 23 creating+peering, 35 unknown, 71 active+clean; 450 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 8.5 KiB/s rd, 2.2 KiB/s wr, 13 op/s 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: cluster 2026-03-07T06:57:45.110683+0000 mgr.vm07.yrfcuj (mgr.14201) 175 : cluster [DBG] pgmap v118: 129 pgs: 23 creating+peering, 35 unknown, 71 active+clean; 450 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 8.5 KiB/s rd, 2.2 KiB/s wr, 13 op/s 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.207807+0000 mon.vm07 (mon.0) 741 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.207807+0000 mon.vm07 (mon.0) 741 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.207887+0000 mon.vm07 (mon.0) 742 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.207887+0000 mon.vm07 (mon.0) 742 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.207988+0000 mon.vm07 (mon.0) 743 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.207988+0000 mon.vm07 (mon.0) 743 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.208033+0000 mon.vm07 (mon.0) 744 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.208033+0000 mon.vm07 (mon.0) 744 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: cluster 2026-03-07T06:57:45.212058+0000 mon.vm07 (mon.0) 745 : cluster [DBG] osdmap e31: 8 total, 8 up, 8 in 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: cluster 2026-03-07T06:57:45.212058+0000 mon.vm07 (mon.0) 745 : cluster [DBG] osdmap e31: 8 total, 8 up, 8 in 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.218554+0000 mon.vm07 (mon.0) 746 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.218554+0000 mon.vm07 (mon.0) 746 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.218647+0000 mon.vm07 (mon.0) 747 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.218647+0000 mon.vm07 (mon.0) 747 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.223515+0000 mon.vm09 (mon.1) 28 : audit [INF] from='client.? 192.168.123.109:0/1001165862' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.223515+0000 mon.vm09 (mon.1) 28 : audit [INF] from='client.? 192.168.123.109:0/1001165862' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.229994+0000 mon.vm07 (mon.0) 748 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.229994+0000 mon.vm07 (mon.0) 748 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.230169+0000 mon.vm07 (mon.0) 749 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.230169+0000 mon.vm07 (mon.0) 749 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.334 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.235041+0000 mon.vm09 (mon.1) 29 : audit [INF] from='client.? 192.168.123.109:0/357189846' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.335 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:45 vm07 bash[17031]: audit 2026-03-07T06:57:45.235041+0000 mon.vm09 (mon.1) 29 : audit [INF] from='client.? 192.168.123.109:0/357189846' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.344 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:44.968380+0000 mon.vm07 (mon.0) 738 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:46.344 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:44.968380+0000 mon.vm07 (mon.0) 738 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:46.344 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:44.973840+0000 mon.vm07 (mon.0) 739 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:46.344 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:44.973840+0000 mon.vm07 (mon.0) 739 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:46.344 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:44.978192+0000 mon.vm07 (mon.0) 740 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:46.344 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:44.978192+0000 mon.vm07 (mon.0) 740 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:46.344 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: cephadm 2026-03-07T06:57:44.979654+0000 mgr.vm07.yrfcuj (mgr.14201) 174 : cephadm [INF] Deploying daemon haproxy.rgw.foo.vm07.rzyepz on vm07 2026-03-07T07:57:46.344 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: cephadm 2026-03-07T06:57:44.979654+0000 mgr.vm07.yrfcuj (mgr.14201) 174 : cephadm [INF] Deploying daemon haproxy.rgw.foo.vm07.rzyepz on vm07 2026-03-07T07:57:46.344 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: cluster 2026-03-07T06:57:45.110683+0000 mgr.vm07.yrfcuj (mgr.14201) 175 : cluster [DBG] pgmap v118: 129 pgs: 23 creating+peering, 35 unknown, 71 active+clean; 450 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 8.5 KiB/s rd, 2.2 KiB/s wr, 13 op/s 2026-03-07T07:57:46.344 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: cluster 2026-03-07T06:57:45.110683+0000 mgr.vm07.yrfcuj (mgr.14201) 175 : cluster [DBG] pgmap v118: 129 pgs: 23 creating+peering, 35 unknown, 71 active+clean; 450 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 8.5 KiB/s rd, 2.2 KiB/s wr, 13 op/s 2026-03-07T07:57:46.344 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.207807+0000 mon.vm07 (mon.0) 741 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-07T07:57:46.344 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.207807+0000 mon.vm07 (mon.0) 741 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-07T07:57:46.344 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.207887+0000 mon.vm07 (mon.0) 742 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-07T07:57:46.344 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.207887+0000 mon.vm07 (mon.0) 742 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-07T07:57:46.344 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.207988+0000 mon.vm07 (mon.0) 743 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-07T07:57:46.344 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.207988+0000 mon.vm07 (mon.0) 743 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-07T07:57:46.345 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.208033+0000 mon.vm07 (mon.0) 744 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-07T07:57:46.345 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.208033+0000 mon.vm07 (mon.0) 744 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-07T07:57:46.345 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: cluster 2026-03-07T06:57:45.212058+0000 mon.vm07 (mon.0) 745 : cluster [DBG] osdmap e31: 8 total, 8 up, 8 in 2026-03-07T07:57:46.345 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: cluster 2026-03-07T06:57:45.212058+0000 mon.vm07 (mon.0) 745 : cluster [DBG] osdmap e31: 8 total, 8 up, 8 in 2026-03-07T07:57:46.345 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.218554+0000 mon.vm07 (mon.0) 746 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.345 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.218554+0000 mon.vm07 (mon.0) 746 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.345 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.218647+0000 mon.vm07 (mon.0) 747 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.345 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.218647+0000 mon.vm07 (mon.0) 747 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.345 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.223515+0000 mon.vm09 (mon.1) 28 : audit [INF] from='client.? 192.168.123.109:0/1001165862' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.345 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.223515+0000 mon.vm09 (mon.1) 28 : audit [INF] from='client.? 192.168.123.109:0/1001165862' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.345 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.229994+0000 mon.vm07 (mon.0) 748 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.345 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.229994+0000 mon.vm07 (mon.0) 748 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.345 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.230169+0000 mon.vm07 (mon.0) 749 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.345 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.230169+0000 mon.vm07 (mon.0) 749 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.345 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.235041+0000 mon.vm09 (mon.1) 29 : audit [INF] from='client.? 192.168.123.109:0/357189846' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:46.345 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:45 vm09 bash[21220]: audit 2026-03-07T06:57:45.235041+0000 mon.vm09 (mon.1) 29 : audit [INF] from='client.? 192.168.123.109:0/357189846' entity='client.rgw.foo.vm09.fgzozy' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-07T07:57:47.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:47 vm09 bash[21220]: audit 2026-03-07T06:57:46.210940+0000 mon.vm07 (mon.0) 750 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-07T07:57:47.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:47 vm09 bash[21220]: audit 2026-03-07T06:57:46.210940+0000 mon.vm07 (mon.0) 750 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-07T07:57:47.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:47 vm09 bash[21220]: audit 2026-03-07T06:57:46.210982+0000 mon.vm07 (mon.0) 751 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-07T07:57:47.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:47 vm09 bash[21220]: audit 2026-03-07T06:57:46.210982+0000 mon.vm07 (mon.0) 751 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-07T07:57:47.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:47 vm09 bash[21220]: audit 2026-03-07T06:57:46.211009+0000 mon.vm07 (mon.0) 752 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-07T07:57:47.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:47 vm09 bash[21220]: audit 2026-03-07T06:57:46.211009+0000 mon.vm07 (mon.0) 752 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-07T07:57:47.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:47 vm09 bash[21220]: audit 2026-03-07T06:57:46.211029+0000 mon.vm07 (mon.0) 753 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-07T07:57:47.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:47 vm09 bash[21220]: audit 2026-03-07T06:57:46.211029+0000 mon.vm07 (mon.0) 753 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-07T07:57:47.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:47 vm09 bash[21220]: cluster 2026-03-07T06:57:46.213872+0000 mon.vm07 (mon.0) 754 : cluster [DBG] osdmap e32: 8 total, 8 up, 8 in 2026-03-07T07:57:47.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:47 vm09 bash[21220]: cluster 2026-03-07T06:57:46.213872+0000 mon.vm07 (mon.0) 754 : cluster [DBG] osdmap e32: 8 total, 8 up, 8 in 2026-03-07T07:57:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:47 vm07 bash[17031]: audit 2026-03-07T06:57:46.210940+0000 mon.vm07 (mon.0) 750 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-07T07:57:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:47 vm07 bash[17031]: audit 2026-03-07T06:57:46.210940+0000 mon.vm07 (mon.0) 750 : audit [INF] from='client.? 192.168.123.107:0/2178684764' entity='client.rgw.foo.vm07.lrxyjy' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-07T07:57:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:47 vm07 bash[17031]: audit 2026-03-07T06:57:46.210982+0000 mon.vm07 (mon.0) 751 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-07T07:57:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:47 vm07 bash[17031]: audit 2026-03-07T06:57:46.210982+0000 mon.vm07 (mon.0) 751 : audit [INF] from='client.? 192.168.123.107:0/624976532' entity='client.rgw.foo.vm07.bgfaka' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-07T07:57:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:47 vm07 bash[17031]: audit 2026-03-07T06:57:46.211009+0000 mon.vm07 (mon.0) 752 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-07T07:57:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:47 vm07 bash[17031]: audit 2026-03-07T06:57:46.211009+0000 mon.vm07 (mon.0) 752 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.kpgoql' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-07T07:57:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:47 vm07 bash[17031]: audit 2026-03-07T06:57:46.211029+0000 mon.vm07 (mon.0) 753 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-07T07:57:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:47 vm07 bash[17031]: audit 2026-03-07T06:57:46.211029+0000 mon.vm07 (mon.0) 753 : audit [INF] from='client.? ' entity='client.rgw.foo.vm09.fgzozy' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-07T07:57:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:47 vm07 bash[17031]: cluster 2026-03-07T06:57:46.213872+0000 mon.vm07 (mon.0) 754 : cluster [DBG] osdmap e32: 8 total, 8 up, 8 in 2026-03-07T07:57:47.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:47 vm07 bash[17031]: cluster 2026-03-07T06:57:46.213872+0000 mon.vm07 (mon.0) 754 : cluster [DBG] osdmap e32: 8 total, 8 up, 8 in 2026-03-07T07:57:48.316 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:57:48.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:48 vm09 bash[21220]: cluster 2026-03-07T06:57:47.152704+0000 mgr.vm07.yrfcuj (mgr.14201) 176 : cluster [DBG] pgmap v121: 129 pgs: 23 creating+peering, 14 unknown, 92 active+clean; 451 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 49 KiB/s rd, 4.0 KiB/s wr, 94 op/s 2026-03-07T07:57:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:48 vm09 bash[21220]: cluster 2026-03-07T06:57:47.152704+0000 mgr.vm07.yrfcuj (mgr.14201) 176 : cluster [DBG] pgmap v121: 129 pgs: 23 creating+peering, 14 unknown, 92 active+clean; 451 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 49 KiB/s rd, 4.0 KiB/s wr, 94 op/s 2026-03-07T07:57:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:48 vm09 bash[21220]: cluster 2026-03-07T06:57:47.210975+0000 mon.vm07 (mon.0) 755 : cluster [INF] Health check cleared: POOL_APP_NOT_ENABLED (was: 1 pool(s) do not have an application enabled) 2026-03-07T07:57:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:48 vm09 bash[21220]: cluster 2026-03-07T06:57:47.210975+0000 mon.vm07 (mon.0) 755 : cluster [INF] Health check cleared: POOL_APP_NOT_ENABLED (was: 1 pool(s) do not have an application enabled) 2026-03-07T07:57:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:48 vm09 bash[21220]: cluster 2026-03-07T06:57:47.210992+0000 mon.vm07 (mon.0) 756 : cluster [INF] Cluster is now healthy 2026-03-07T07:57:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:48 vm09 bash[21220]: cluster 2026-03-07T06:57:47.210992+0000 mon.vm07 (mon.0) 756 : cluster [INF] Cluster is now healthy 2026-03-07T07:57:48.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:48 vm07 bash[17031]: cluster 2026-03-07T06:57:47.152704+0000 mgr.vm07.yrfcuj (mgr.14201) 176 : cluster [DBG] pgmap v121: 129 pgs: 23 creating+peering, 14 unknown, 92 active+clean; 451 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 49 KiB/s rd, 4.0 KiB/s wr, 94 op/s 2026-03-07T07:57:48.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:48 vm07 bash[17031]: cluster 2026-03-07T06:57:47.152704+0000 mgr.vm07.yrfcuj (mgr.14201) 176 : cluster [DBG] pgmap v121: 129 pgs: 23 creating+peering, 14 unknown, 92 active+clean; 451 KiB data, 213 MiB used, 160 GiB / 160 GiB avail; 49 KiB/s rd, 4.0 KiB/s wr, 94 op/s 2026-03-07T07:57:48.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:48 vm07 bash[17031]: cluster 2026-03-07T06:57:47.210975+0000 mon.vm07 (mon.0) 755 : cluster [INF] Health check cleared: POOL_APP_NOT_ENABLED (was: 1 pool(s) do not have an application enabled) 2026-03-07T07:57:48.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:48 vm07 bash[17031]: cluster 2026-03-07T06:57:47.210975+0000 mon.vm07 (mon.0) 755 : cluster [INF] Health check cleared: POOL_APP_NOT_ENABLED (was: 1 pool(s) do not have an application enabled) 2026-03-07T07:57:48.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:48 vm07 bash[17031]: cluster 2026-03-07T06:57:47.210992+0000 mon.vm07 (mon.0) 756 : cluster [INF] Cluster is now healthy 2026-03-07T07:57:48.659 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:48 vm07 bash[17031]: cluster 2026-03-07T06:57:47.210992+0000 mon.vm07 (mon.0) 756 : cluster [INF] Cluster is now healthy 2026-03-07T07:57:49.162 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:57:49.162 INFO:teuthology.orchestra.run.vm07.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-07T06:53:25.695727Z", "last_refresh": "2026-03-07T06:57:30.067219Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-07T06:54:32.927755Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-07T06:53:24.247577Z", "last_refresh": "2026-03-07T06:57:29.595811Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:33.727624Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-07T06:53:23.837568Z", "last_refresh": "2026-03-07T06:57:29.595876Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-07T06:53:24.987960Z", "last_refresh": "2026-03-07T06:57:30.067033Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-07T06:57:44.978406Z service:ingress.rgw.foo [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "rgw.foo", "service_name": "ingress.rgw.foo", "service_type": "ingress", "spec": {"backend_service": "rgw.foo", "first_virtual_router_id": 50, "frontend_port": 9000, "monitor_port": 9001, "virtual_ip": "12.12.1.107/22"}, "status": {"created": "2026-03-07T06:57:29.846008Z", "ports": [9000, 9001], "running": 0, "size": 4, "virtual_ip": "12.12.1.107/22"}}, {"events": ["2026-03-07T06:54:35.185395Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-07T06:53:23.396871Z", "last_refresh": "2026-03-07T06:57:29.595909Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:36.226821Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm07:192.168.123.107=vm07", "vm09:192.168.123.109=vm09"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-07T06:54:13.493474Z", "last_refresh": "2026-03-07T06:57:29.595844Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:34.440095Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-07T06:53:25.333423Z", "last_refresh": "2026-03-07T06:57:29.596000Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:54.121213Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-07T06:54:54.116553Z", "last_refresh": "2026-03-07T06:57:29.595721Z", "running": 8, "size": 8}}, {"events": ["2026-03-07T06:54:36.229563Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-07T06:53:24.610265Z", "last_refresh": "2026-03-07T06:57:30.066979Z", "ports": [9095], "running": 1, "size": 1}}, {"events": ["2026-03-07T06:57:39.780378Z service:rgw.foo [INFO] \"service was created\""], "placement": {"count": 4, "host_pattern": "*"}, "service_id": "foo", "service_name": "rgw.foo", "service_type": "rgw", "spec": {"rgw_frontend_port": 8000}, "status": {"created": "2026-03-07T06:57:39.774719Z", "ports": [8000], "running": 0, "size": 4}}] 2026-03-07T07:57:49.251 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:49 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:57:49.307 INFO:tasks.cephadm:rgw.foo has 0/4 2026-03-07T07:57:49.537 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:49 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:57:50.309 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph orch ls -f json 2026-03-07T07:57:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:50 vm07 bash[17031]: cluster 2026-03-07T06:57:49.153104+0000 mgr.vm07.yrfcuj (mgr.14201) 177 : cluster [DBG] pgmap v122: 129 pgs: 5 creating+peering, 124 active+clean; 453 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 211 KiB/s rd, 5.8 KiB/s wr, 371 op/s 2026-03-07T07:57:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:50 vm07 bash[17031]: cluster 2026-03-07T06:57:49.153104+0000 mgr.vm07.yrfcuj (mgr.14201) 177 : cluster [DBG] pgmap v122: 129 pgs: 5 creating+peering, 124 active+clean; 453 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 211 KiB/s rd, 5.8 KiB/s wr, 371 op/s 2026-03-07T07:57:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:50 vm07 bash[17031]: audit 2026-03-07T06:57:49.154588+0000 mgr.vm07.yrfcuj (mgr.14201) 178 : audit [DBG] from='client.24351 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:57:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:50 vm07 bash[17031]: audit 2026-03-07T06:57:49.154588+0000 mgr.vm07.yrfcuj (mgr.14201) 178 : audit [DBG] from='client.24351 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:57:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:50 vm07 bash[17031]: audit 2026-03-07T06:57:49.477139+0000 mon.vm07 (mon.0) 757 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:50 vm07 bash[17031]: audit 2026-03-07T06:57:49.477139+0000 mon.vm07 (mon.0) 757 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:50 vm07 bash[17031]: audit 2026-03-07T06:57:49.483422+0000 mon.vm07 (mon.0) 758 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:50 vm07 bash[17031]: audit 2026-03-07T06:57:49.483422+0000 mon.vm07 (mon.0) 758 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:50 vm07 bash[17031]: audit 2026-03-07T06:57:49.493059+0000 mon.vm07 (mon.0) 759 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:50 vm07 bash[17031]: audit 2026-03-07T06:57:49.493059+0000 mon.vm07 (mon.0) 759 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:50 vm07 bash[17031]: audit 2026-03-07T06:57:49.501725+0000 mon.vm07 (mon.0) 760 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:50 vm07 bash[17031]: audit 2026-03-07T06:57:49.501725+0000 mon.vm07 (mon.0) 760 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:50 vm07 bash[17031]: cephadm 2026-03-07T06:57:49.502096+0000 mgr.vm07.yrfcuj (mgr.14201) 179 : cephadm [INF] 12.12.1.107 is in 12.12.0.0/22 on vm07 interface ens3 2026-03-07T07:57:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:50 vm07 bash[17031]: cephadm 2026-03-07T06:57:49.502096+0000 mgr.vm07.yrfcuj (mgr.14201) 179 : cephadm [INF] 12.12.1.107 is in 12.12.0.0/22 on vm07 interface ens3 2026-03-07T07:57:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:50 vm07 bash[17031]: cephadm 2026-03-07T06:57:49.502138+0000 mgr.vm07.yrfcuj (mgr.14201) 180 : cephadm [INF] 12.12.1.107 is in 12.12.0.0/22 on vm09 interface ens3 2026-03-07T07:57:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:50 vm07 bash[17031]: cephadm 2026-03-07T06:57:49.502138+0000 mgr.vm07.yrfcuj (mgr.14201) 180 : cephadm [INF] 12.12.1.107 is in 12.12.0.0/22 on vm09 interface ens3 2026-03-07T07:57:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:50 vm07 bash[17031]: cephadm 2026-03-07T06:57:49.504519+0000 mgr.vm07.yrfcuj (mgr.14201) 181 : cephadm [INF] Deploying daemon keepalived.rgw.foo.vm07.mbvpkm on vm07 2026-03-07T07:57:50.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:50 vm07 bash[17031]: cephadm 2026-03-07T06:57:49.504519+0000 mgr.vm07.yrfcuj (mgr.14201) 181 : cephadm [INF] Deploying daemon keepalived.rgw.foo.vm07.mbvpkm on vm07 2026-03-07T07:57:50.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:50 vm09 bash[21220]: cluster 2026-03-07T06:57:49.153104+0000 mgr.vm07.yrfcuj (mgr.14201) 177 : cluster [DBG] pgmap v122: 129 pgs: 5 creating+peering, 124 active+clean; 453 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 211 KiB/s rd, 5.8 KiB/s wr, 371 op/s 2026-03-07T07:57:50.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:50 vm09 bash[21220]: cluster 2026-03-07T06:57:49.153104+0000 mgr.vm07.yrfcuj (mgr.14201) 177 : cluster [DBG] pgmap v122: 129 pgs: 5 creating+peering, 124 active+clean; 453 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 211 KiB/s rd, 5.8 KiB/s wr, 371 op/s 2026-03-07T07:57:50.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:50 vm09 bash[21220]: audit 2026-03-07T06:57:49.154588+0000 mgr.vm07.yrfcuj (mgr.14201) 178 : audit [DBG] from='client.24351 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:57:50.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:50 vm09 bash[21220]: audit 2026-03-07T06:57:49.154588+0000 mgr.vm07.yrfcuj (mgr.14201) 178 : audit [DBG] from='client.24351 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:57:50.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:50 vm09 bash[21220]: audit 2026-03-07T06:57:49.477139+0000 mon.vm07 (mon.0) 757 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:50.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:50 vm09 bash[21220]: audit 2026-03-07T06:57:49.477139+0000 mon.vm07 (mon.0) 757 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:50.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:50 vm09 bash[21220]: audit 2026-03-07T06:57:49.483422+0000 mon.vm07 (mon.0) 758 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:50.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:50 vm09 bash[21220]: audit 2026-03-07T06:57:49.483422+0000 mon.vm07 (mon.0) 758 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:50 vm09 bash[21220]: audit 2026-03-07T06:57:49.493059+0000 mon.vm07 (mon.0) 759 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:50 vm09 bash[21220]: audit 2026-03-07T06:57:49.493059+0000 mon.vm07 (mon.0) 759 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:50 vm09 bash[21220]: audit 2026-03-07T06:57:49.501725+0000 mon.vm07 (mon.0) 760 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:50 vm09 bash[21220]: audit 2026-03-07T06:57:49.501725+0000 mon.vm07 (mon.0) 760 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:50 vm09 bash[21220]: cephadm 2026-03-07T06:57:49.502096+0000 mgr.vm07.yrfcuj (mgr.14201) 179 : cephadm [INF] 12.12.1.107 is in 12.12.0.0/22 on vm07 interface ens3 2026-03-07T07:57:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:50 vm09 bash[21220]: cephadm 2026-03-07T06:57:49.502096+0000 mgr.vm07.yrfcuj (mgr.14201) 179 : cephadm [INF] 12.12.1.107 is in 12.12.0.0/22 on vm07 interface ens3 2026-03-07T07:57:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:50 vm09 bash[21220]: cephadm 2026-03-07T06:57:49.502138+0000 mgr.vm07.yrfcuj (mgr.14201) 180 : cephadm [INF] 12.12.1.107 is in 12.12.0.0/22 on vm09 interface ens3 2026-03-07T07:57:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:50 vm09 bash[21220]: cephadm 2026-03-07T06:57:49.502138+0000 mgr.vm07.yrfcuj (mgr.14201) 180 : cephadm [INF] 12.12.1.107 is in 12.12.0.0/22 on vm09 interface ens3 2026-03-07T07:57:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:50 vm09 bash[21220]: cephadm 2026-03-07T06:57:49.504519+0000 mgr.vm07.yrfcuj (mgr.14201) 181 : cephadm [INF] Deploying daemon keepalived.rgw.foo.vm07.mbvpkm on vm07 2026-03-07T07:57:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:50 vm09 bash[21220]: cephadm 2026-03-07T06:57:49.504519+0000 mgr.vm07.yrfcuj (mgr.14201) 181 : cephadm [INF] Deploying daemon keepalived.rgw.foo.vm07.mbvpkm on vm07 2026-03-07T07:57:52.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:52 vm09 bash[21220]: cluster 2026-03-07T06:57:51.153460+0000 mgr.vm07.yrfcuj (mgr.14201) 182 : cluster [DBG] pgmap v123: 129 pgs: 129 active+clean; 454 KiB data, 222 MiB used, 160 GiB / 160 GiB avail; 220 KiB/s rd, 4.9 KiB/s wr, 388 op/s 2026-03-07T07:57:52.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:52 vm09 bash[21220]: cluster 2026-03-07T06:57:51.153460+0000 mgr.vm07.yrfcuj (mgr.14201) 182 : cluster [DBG] pgmap v123: 129 pgs: 129 active+clean; 454 KiB data, 222 MiB used, 160 GiB / 160 GiB avail; 220 KiB/s rd, 4.9 KiB/s wr, 388 op/s 2026-03-07T07:57:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:52 vm07 bash[17031]: cluster 2026-03-07T06:57:51.153460+0000 mgr.vm07.yrfcuj (mgr.14201) 182 : cluster [DBG] pgmap v123: 129 pgs: 129 active+clean; 454 KiB data, 222 MiB used, 160 GiB / 160 GiB avail; 220 KiB/s rd, 4.9 KiB/s wr, 388 op/s 2026-03-07T07:57:52.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:52 vm07 bash[17031]: cluster 2026-03-07T06:57:51.153460+0000 mgr.vm07.yrfcuj (mgr.14201) 182 : cluster [DBG] pgmap v123: 129 pgs: 129 active+clean; 454 KiB data, 222 MiB used, 160 GiB / 160 GiB avail; 220 KiB/s rd, 4.9 KiB/s wr, 388 op/s 2026-03-07T07:57:53.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:53 vm09 bash[21220]: cluster 2026-03-07T06:57:53.155193+0000 mgr.vm07.yrfcuj (mgr.14201) 183 : cluster [DBG] pgmap v124: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 260 KiB/s rd, 4.2 KiB/s wr, 454 op/s 2026-03-07T07:57:53.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:53 vm09 bash[21220]: cluster 2026-03-07T06:57:53.155193+0000 mgr.vm07.yrfcuj (mgr.14201) 183 : cluster [DBG] pgmap v124: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 260 KiB/s rd, 4.2 KiB/s wr, 454 op/s 2026-03-07T07:57:53.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:53 vm07 bash[17031]: cluster 2026-03-07T06:57:53.155193+0000 mgr.vm07.yrfcuj (mgr.14201) 183 : cluster [DBG] pgmap v124: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 260 KiB/s rd, 4.2 KiB/s wr, 454 op/s 2026-03-07T07:57:53.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:53 vm07 bash[17031]: cluster 2026-03-07T06:57:53.155193+0000 mgr.vm07.yrfcuj (mgr.14201) 183 : cluster [DBG] pgmap v124: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 260 KiB/s rd, 4.2 KiB/s wr, 454 op/s 2026-03-07T07:57:54.908 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:54 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:57:55.104 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:57:55.216 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:55 vm07 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:57:55.564 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:57:55.564 INFO:teuthology.orchestra.run.vm07.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-07T06:53:25.695727Z", "last_refresh": "2026-03-07T06:57:30.067219Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-07T06:54:32.927755Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-07T06:53:24.247577Z", "last_refresh": "2026-03-07T06:57:29.595811Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:33.727624Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-07T06:53:23.837568Z", "last_refresh": "2026-03-07T06:57:29.595876Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-07T06:53:24.987960Z", "last_refresh": "2026-03-07T06:57:30.067033Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-07T06:57:55.204863Z service:ingress.rgw.foo [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "rgw.foo", "service_name": "ingress.rgw.foo", "service_type": "ingress", "spec": {"backend_service": "rgw.foo", "first_virtual_router_id": 50, "frontend_port": 9000, "monitor_port": 9001, "virtual_ip": "12.12.1.107/22"}, "status": {"created": "2026-03-07T06:57:29.846008Z", "ports": [9000, 9001], "running": 0, "size": 4, "virtual_ip": "12.12.1.107/22"}}, {"events": ["2026-03-07T06:54:35.185395Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-07T06:53:23.396871Z", "last_refresh": "2026-03-07T06:57:29.595909Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:36.226821Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm07:192.168.123.107=vm07", "vm09:192.168.123.109=vm09"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-07T06:54:13.493474Z", "last_refresh": "2026-03-07T06:57:29.595844Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:34.440095Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-07T06:53:25.333423Z", "last_refresh": "2026-03-07T06:57:29.596000Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:54.121213Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-07T06:54:54.116553Z", "last_refresh": "2026-03-07T06:57:29.595721Z", "running": 8, "size": 8}}, {"events": ["2026-03-07T06:54:36.229563Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-07T06:53:24.610265Z", "last_refresh": "2026-03-07T06:57:30.066979Z", "ports": [9095], "running": 1, "size": 1}}, {"events": ["2026-03-07T06:57:39.780378Z service:rgw.foo [INFO] \"service was created\""], "placement": {"count": 4, "host_pattern": "*"}, "service_id": "foo", "service_name": "rgw.foo", "service_type": "rgw", "spec": {"rgw_frontend_port": 8000}, "status": {"created": "2026-03-07T06:57:39.774719Z", "ports": [8000], "running": 0, "size": 4}}] 2026-03-07T07:57:55.628 INFO:tasks.cephadm:rgw.foo has 0/4 2026-03-07T07:57:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:56 vm07 bash[17031]: audit 2026-03-07T06:57:55.131044+0000 mon.vm07 (mon.0) 761 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:56 vm07 bash[17031]: audit 2026-03-07T06:57:55.131044+0000 mon.vm07 (mon.0) 761 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:56 vm07 bash[17031]: audit 2026-03-07T06:57:55.131770+0000 mon.vm07 (mon.0) 762 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:57:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:56 vm07 bash[17031]: audit 2026-03-07T06:57:55.131770+0000 mon.vm07 (mon.0) 762 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:57:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:56 vm07 bash[17031]: cluster 2026-03-07T06:57:55.158826+0000 mgr.vm07.yrfcuj (mgr.14201) 184 : cluster [DBG] pgmap v125: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 210 KiB/s rd, 3.4 KiB/s wr, 367 op/s 2026-03-07T07:57:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:56 vm07 bash[17031]: cluster 2026-03-07T06:57:55.158826+0000 mgr.vm07.yrfcuj (mgr.14201) 184 : cluster [DBG] pgmap v125: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 210 KiB/s rd, 3.4 KiB/s wr, 367 op/s 2026-03-07T07:57:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:56 vm07 bash[17031]: audit 2026-03-07T06:57:55.194984+0000 mon.vm07 (mon.0) 763 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:56 vm07 bash[17031]: audit 2026-03-07T06:57:55.194984+0000 mon.vm07 (mon.0) 763 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:56 vm07 bash[17031]: audit 2026-03-07T06:57:55.200093+0000 mon.vm07 (mon.0) 764 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:56 vm07 bash[17031]: audit 2026-03-07T06:57:55.200093+0000 mon.vm07 (mon.0) 764 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:56 vm07 bash[17031]: audit 2026-03-07T06:57:55.204535+0000 mon.vm07 (mon.0) 765 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:56 vm07 bash[17031]: audit 2026-03-07T06:57:55.204535+0000 mon.vm07 (mon.0) 765 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:56 vm07 bash[17031]: cephadm 2026-03-07T06:57:55.205719+0000 mgr.vm07.yrfcuj (mgr.14201) 185 : cephadm [INF] 12.12.1.107 is in 12.12.0.0/22 on vm09 interface ens3 2026-03-07T07:57:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:56 vm07 bash[17031]: cephadm 2026-03-07T06:57:55.205719+0000 mgr.vm07.yrfcuj (mgr.14201) 185 : cephadm [INF] 12.12.1.107 is in 12.12.0.0/22 on vm09 interface ens3 2026-03-07T07:57:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:56 vm07 bash[17031]: cephadm 2026-03-07T06:57:55.205759+0000 mgr.vm07.yrfcuj (mgr.14201) 186 : cephadm [INF] 12.12.1.107 is in 12.12.0.0/22 on vm07 interface ens3 2026-03-07T07:57:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:56 vm07 bash[17031]: cephadm 2026-03-07T06:57:55.205759+0000 mgr.vm07.yrfcuj (mgr.14201) 186 : cephadm [INF] 12.12.1.107 is in 12.12.0.0/22 on vm07 interface ens3 2026-03-07T07:57:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:56 vm07 bash[17031]: cephadm 2026-03-07T06:57:55.206044+0000 mgr.vm07.yrfcuj (mgr.14201) 187 : cephadm [INF] Deploying daemon keepalived.rgw.foo.vm09.bnmvqb on vm09 2026-03-07T07:57:56.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:56 vm07 bash[17031]: cephadm 2026-03-07T06:57:55.206044+0000 mgr.vm07.yrfcuj (mgr.14201) 187 : cephadm [INF] Deploying daemon keepalived.rgw.foo.vm09.bnmvqb on vm09 2026-03-07T07:57:56.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:56 vm09 bash[21220]: audit 2026-03-07T06:57:55.131044+0000 mon.vm07 (mon.0) 761 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:56.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:56 vm09 bash[21220]: audit 2026-03-07T06:57:55.131044+0000 mon.vm07 (mon.0) 761 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:56.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:56 vm09 bash[21220]: audit 2026-03-07T06:57:55.131770+0000 mon.vm07 (mon.0) 762 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:57:56.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:56 vm09 bash[21220]: audit 2026-03-07T06:57:55.131770+0000 mon.vm07 (mon.0) 762 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:57:56.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:56 vm09 bash[21220]: cluster 2026-03-07T06:57:55.158826+0000 mgr.vm07.yrfcuj (mgr.14201) 184 : cluster [DBG] pgmap v125: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 210 KiB/s rd, 3.4 KiB/s wr, 367 op/s 2026-03-07T07:57:56.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:56 vm09 bash[21220]: cluster 2026-03-07T06:57:55.158826+0000 mgr.vm07.yrfcuj (mgr.14201) 184 : cluster [DBG] pgmap v125: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 210 KiB/s rd, 3.4 KiB/s wr, 367 op/s 2026-03-07T07:57:56.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:56 vm09 bash[21220]: audit 2026-03-07T06:57:55.194984+0000 mon.vm07 (mon.0) 763 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:56.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:56 vm09 bash[21220]: audit 2026-03-07T06:57:55.194984+0000 mon.vm07 (mon.0) 763 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:56.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:56 vm09 bash[21220]: audit 2026-03-07T06:57:55.200093+0000 mon.vm07 (mon.0) 764 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:56.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:56 vm09 bash[21220]: audit 2026-03-07T06:57:55.200093+0000 mon.vm07 (mon.0) 764 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:56 vm09 bash[21220]: audit 2026-03-07T06:57:55.204535+0000 mon.vm07 (mon.0) 765 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:56 vm09 bash[21220]: audit 2026-03-07T06:57:55.204535+0000 mon.vm07 (mon.0) 765 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:56 vm09 bash[21220]: cephadm 2026-03-07T06:57:55.205719+0000 mgr.vm07.yrfcuj (mgr.14201) 185 : cephadm [INF] 12.12.1.107 is in 12.12.0.0/22 on vm09 interface ens3 2026-03-07T07:57:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:56 vm09 bash[21220]: cephadm 2026-03-07T06:57:55.205719+0000 mgr.vm07.yrfcuj (mgr.14201) 185 : cephadm [INF] 12.12.1.107 is in 12.12.0.0/22 on vm09 interface ens3 2026-03-07T07:57:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:56 vm09 bash[21220]: cephadm 2026-03-07T06:57:55.205759+0000 mgr.vm07.yrfcuj (mgr.14201) 186 : cephadm [INF] 12.12.1.107 is in 12.12.0.0/22 on vm07 interface ens3 2026-03-07T07:57:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:56 vm09 bash[21220]: cephadm 2026-03-07T06:57:55.205759+0000 mgr.vm07.yrfcuj (mgr.14201) 186 : cephadm [INF] 12.12.1.107 is in 12.12.0.0/22 on vm07 interface ens3 2026-03-07T07:57:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:56 vm09 bash[21220]: cephadm 2026-03-07T06:57:55.206044+0000 mgr.vm07.yrfcuj (mgr.14201) 187 : cephadm [INF] Deploying daemon keepalived.rgw.foo.vm09.bnmvqb on vm09 2026-03-07T07:57:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:56 vm09 bash[21220]: cephadm 2026-03-07T06:57:55.206044+0000 mgr.vm07.yrfcuj (mgr.14201) 187 : cephadm [INF] Deploying daemon keepalived.rgw.foo.vm09.bnmvqb on vm09 2026-03-07T07:57:56.629 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph orch ls -f json 2026-03-07T07:57:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:57 vm07 bash[17031]: audit 2026-03-07T06:57:55.557035+0000 mgr.vm07.yrfcuj (mgr.14201) 188 : audit [DBG] from='client.14546 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:57:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:57 vm07 bash[17031]: audit 2026-03-07T06:57:55.557035+0000 mgr.vm07.yrfcuj (mgr.14201) 188 : audit [DBG] from='client.14546 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:57:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:57 vm07 bash[17031]: audit 2026-03-07T06:57:56.644234+0000 mon.vm07 (mon.0) 766 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:57.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:57 vm07 bash[17031]: audit 2026-03-07T06:57:56.644234+0000 mon.vm07 (mon.0) 766 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:57.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:57 vm09 bash[21220]: audit 2026-03-07T06:57:55.557035+0000 mgr.vm07.yrfcuj (mgr.14201) 188 : audit [DBG] from='client.14546 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:57:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:57 vm09 bash[21220]: audit 2026-03-07T06:57:55.557035+0000 mgr.vm07.yrfcuj (mgr.14201) 188 : audit [DBG] from='client.14546 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:57:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:57 vm09 bash[21220]: audit 2026-03-07T06:57:56.644234+0000 mon.vm07 (mon.0) 766 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:57 vm09 bash[21220]: audit 2026-03-07T06:57:56.644234+0000 mon.vm07 (mon.0) 766 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:57:58.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:58 vm09 bash[21220]: cluster 2026-03-07T06:57:57.159120+0000 mgr.vm07.yrfcuj (mgr.14201) 189 : cluster [DBG] pgmap v126: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 191 KiB/s rd, 3.1 KiB/s wr, 333 op/s 2026-03-07T07:57:58.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:57:58 vm09 bash[21220]: cluster 2026-03-07T06:57:57.159120+0000 mgr.vm07.yrfcuj (mgr.14201) 189 : cluster [DBG] pgmap v126: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 191 KiB/s rd, 3.1 KiB/s wr, 333 op/s 2026-03-07T07:57:58.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:58 vm07 bash[17031]: cluster 2026-03-07T06:57:57.159120+0000 mgr.vm07.yrfcuj (mgr.14201) 189 : cluster [DBG] pgmap v126: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 191 KiB/s rd, 3.1 KiB/s wr, 333 op/s 2026-03-07T07:57:58.408 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:57:58 vm07 bash[17031]: cluster 2026-03-07T06:57:57.159120+0000 mgr.vm07.yrfcuj (mgr.14201) 189 : cluster [DBG] pgmap v126: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 191 KiB/s rd, 3.1 KiB/s wr, 333 op/s 2026-03-07T07:58:00.559 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:00 vm09 bash[21220]: cluster 2026-03-07T06:57:59.159523+0000 mgr.vm07.yrfcuj (mgr.14201) 190 : cluster [DBG] pgmap v127: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 159 KiB/s rd, 1.8 KiB/s wr, 274 op/s 2026-03-07T07:58:00.559 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:00 vm09 bash[21220]: cluster 2026-03-07T06:57:59.159523+0000 mgr.vm07.yrfcuj (mgr.14201) 190 : cluster [DBG] pgmap v127: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 159 KiB/s rd, 1.8 KiB/s wr, 274 op/s 2026-03-07T07:58:00.559 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:00 vm09 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:58:00.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:00 vm07 bash[17031]: cluster 2026-03-07T06:57:59.159523+0000 mgr.vm07.yrfcuj (mgr.14201) 190 : cluster [DBG] pgmap v127: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 159 KiB/s rd, 1.8 KiB/s wr, 274 op/s 2026-03-07T07:58:00.658 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:00 vm07 bash[17031]: cluster 2026-03-07T06:57:59.159523+0000 mgr.vm07.yrfcuj (mgr.14201) 190 : cluster [DBG] pgmap v127: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 159 KiB/s rd, 1.8 KiB/s wr, 274 op/s 2026-03-07T07:58:00.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:00 vm09 systemd[1]: /etc/systemd/system/ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-07T07:58:01.439 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:58:01.829 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:58:01.829 INFO:teuthology.orchestra.run.vm07.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-07T06:53:25.695727Z", "last_refresh": "2026-03-07T06:57:30.067219Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-07T06:54:32.927755Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-07T06:53:24.247577Z", "last_refresh": "2026-03-07T06:57:29.595811Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:33.727624Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-07T06:53:23.837568Z", "last_refresh": "2026-03-07T06:57:29.595876Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-07T06:53:24.987960Z", "last_refresh": "2026-03-07T06:57:30.067033Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-07T06:58:00.908906Z service:ingress.rgw.foo [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "rgw.foo", "service_name": "ingress.rgw.foo", "service_type": "ingress", "spec": {"backend_service": "rgw.foo", "first_virtual_router_id": 50, "frontend_port": 9000, "monitor_port": 9001, "virtual_ip": "12.12.1.107/22"}, "status": {"created": "2026-03-07T06:57:29.846008Z", "ports": [9000, 9001], "running": 0, "size": 4, "virtual_ip": "12.12.1.107/22"}}, {"events": ["2026-03-07T06:54:35.185395Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-07T06:53:23.396871Z", "last_refresh": "2026-03-07T06:57:29.595909Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:36.226821Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm07:192.168.123.107=vm07", "vm09:192.168.123.109=vm09"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-07T06:54:13.493474Z", "last_refresh": "2026-03-07T06:57:29.595844Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:34.440095Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-07T06:53:25.333423Z", "last_refresh": "2026-03-07T06:57:29.596000Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:54.121213Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-07T06:54:54.116553Z", "last_refresh": "2026-03-07T06:57:29.595721Z", "running": 8, "size": 8}}, {"events": ["2026-03-07T06:54:36.229563Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-07T06:53:24.610265Z", "last_refresh": "2026-03-07T06:57:30.066979Z", "ports": [9095], "running": 1, "size": 1}}, {"events": ["2026-03-07T06:57:39.780378Z service:rgw.foo [INFO] \"service was created\""], "placement": {"count": 4, "host_pattern": "*"}, "service_id": "foo", "service_name": "rgw.foo", "service_type": "rgw", "spec": {"rgw_frontend_port": 8000}, "status": {"created": "2026-03-07T06:57:39.774719Z", "ports": [8000], "running": 0, "size": 4}}] 2026-03-07T07:58:01.889 INFO:tasks.cephadm:rgw.foo has 0/4 2026-03-07T07:58:02.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:01 vm07 bash[17031]: audit 2026-03-07T06:58:00.891390+0000 mon.vm07 (mon.0) 767 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:01 vm07 bash[17031]: audit 2026-03-07T06:58:00.891390+0000 mon.vm07 (mon.0) 767 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:01 vm07 bash[17031]: audit 2026-03-07T06:58:00.901198+0000 mon.vm07 (mon.0) 768 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:01 vm07 bash[17031]: audit 2026-03-07T06:58:00.901198+0000 mon.vm07 (mon.0) 768 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:01 vm07 bash[17031]: audit 2026-03-07T06:58:00.905931+0000 mon.vm07 (mon.0) 769 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:01 vm07 bash[17031]: audit 2026-03-07T06:58:00.905931+0000 mon.vm07 (mon.0) 769 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:01 vm07 bash[17031]: audit 2026-03-07T06:58:00.908731+0000 mon.vm07 (mon.0) 770 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:01 vm07 bash[17031]: audit 2026-03-07T06:58:00.908731+0000 mon.vm07 (mon.0) 770 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:01 vm07 bash[17031]: audit 2026-03-07T06:58:00.929955+0000 mon.vm07 (mon.0) 771 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:58:02.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:01 vm07 bash[17031]: audit 2026-03-07T06:58:00.929955+0000 mon.vm07 (mon.0) 771 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:58:02.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:01 vm07 bash[17031]: cluster 2026-03-07T06:58:01.159829+0000 mgr.vm07.yrfcuj (mgr.14201) 191 : cluster [DBG] pgmap v128: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 71 KiB/s rd, 255 B/s wr, 121 op/s 2026-03-07T07:58:02.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:01 vm07 bash[17031]: cluster 2026-03-07T06:58:01.159829+0000 mgr.vm07.yrfcuj (mgr.14201) 191 : cluster [DBG] pgmap v128: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 71 KiB/s rd, 255 B/s wr, 121 op/s 2026-03-07T07:58:02.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:01 vm07 bash[17031]: audit 2026-03-07T06:58:01.649621+0000 mon.vm07 (mon.0) 772 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:01 vm07 bash[17031]: audit 2026-03-07T06:58:01.649621+0000 mon.vm07 (mon.0) 772 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.279 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:01 vm09 bash[21220]: audit 2026-03-07T06:58:00.891390+0000 mon.vm07 (mon.0) 767 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.279 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:01 vm09 bash[21220]: audit 2026-03-07T06:58:00.891390+0000 mon.vm07 (mon.0) 767 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.279 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:01 vm09 bash[21220]: audit 2026-03-07T06:58:00.901198+0000 mon.vm07 (mon.0) 768 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.279 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:01 vm09 bash[21220]: audit 2026-03-07T06:58:00.901198+0000 mon.vm07 (mon.0) 768 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.279 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:01 vm09 bash[21220]: audit 2026-03-07T06:58:00.905931+0000 mon.vm07 (mon.0) 769 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.279 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:01 vm09 bash[21220]: audit 2026-03-07T06:58:00.905931+0000 mon.vm07 (mon.0) 769 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.279 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:01 vm09 bash[21220]: audit 2026-03-07T06:58:00.908731+0000 mon.vm07 (mon.0) 770 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.279 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:01 vm09 bash[21220]: audit 2026-03-07T06:58:00.908731+0000 mon.vm07 (mon.0) 770 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.279 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:01 vm09 bash[21220]: audit 2026-03-07T06:58:00.929955+0000 mon.vm07 (mon.0) 771 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:58:02.279 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:01 vm09 bash[21220]: audit 2026-03-07T06:58:00.929955+0000 mon.vm07 (mon.0) 771 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:58:02.279 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:01 vm09 bash[21220]: cluster 2026-03-07T06:58:01.159829+0000 mgr.vm07.yrfcuj (mgr.14201) 191 : cluster [DBG] pgmap v128: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 71 KiB/s rd, 255 B/s wr, 121 op/s 2026-03-07T07:58:02.279 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:01 vm09 bash[21220]: cluster 2026-03-07T06:58:01.159829+0000 mgr.vm07.yrfcuj (mgr.14201) 191 : cluster [DBG] pgmap v128: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 71 KiB/s rd, 255 B/s wr, 121 op/s 2026-03-07T07:58:02.279 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:01 vm09 bash[21220]: audit 2026-03-07T06:58:01.649621+0000 mon.vm07 (mon.0) 772 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.279 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:01 vm09 bash[21220]: audit 2026-03-07T06:58:01.649621+0000 mon.vm07 (mon.0) 772 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:02.890 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph orch ls -f json 2026-03-07T07:58:03.157 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:02 vm07 bash[17031]: audit 2026-03-07T06:58:01.821666+0000 mgr.vm07.yrfcuj (mgr.14201) 192 : audit [DBG] from='client.14550 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:58:03.158 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:02 vm07 bash[17031]: audit 2026-03-07T06:58:01.821666+0000 mgr.vm07.yrfcuj (mgr.14201) 192 : audit [DBG] from='client.14550 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:58:03.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:02 vm09 bash[21220]: audit 2026-03-07T06:58:01.821666+0000 mgr.vm07.yrfcuj (mgr.14201) 192 : audit [DBG] from='client.14550 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:58:03.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:02 vm09 bash[21220]: audit 2026-03-07T06:58:01.821666+0000 mgr.vm07.yrfcuj (mgr.14201) 192 : audit [DBG] from='client.14550 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:58:04.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:03 vm09 bash[21220]: cluster 2026-03-07T06:58:03.160293+0000 mgr.vm07.yrfcuj (mgr.14201) 193 : cluster [DBG] pgmap v129: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 48 KiB/s rd, 0 B/s wr, 79 op/s 2026-03-07T07:58:04.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:03 vm09 bash[21220]: cluster 2026-03-07T06:58:03.160293+0000 mgr.vm07.yrfcuj (mgr.14201) 193 : cluster [DBG] pgmap v129: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 48 KiB/s rd, 0 B/s wr, 79 op/s 2026-03-07T07:58:04.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:03 vm07 bash[17031]: cluster 2026-03-07T06:58:03.160293+0000 mgr.vm07.yrfcuj (mgr.14201) 193 : cluster [DBG] pgmap v129: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 48 KiB/s rd, 0 B/s wr, 79 op/s 2026-03-07T07:58:04.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:03 vm07 bash[17031]: cluster 2026-03-07T06:58:03.160293+0000 mgr.vm07.yrfcuj (mgr.14201) 193 : cluster [DBG] pgmap v129: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail; 48 KiB/s rd, 0 B/s wr, 79 op/s 2026-03-07T07:58:06.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:06 vm09 bash[21220]: cluster 2026-03-07T06:58:05.160637+0000 mgr.vm07.yrfcuj (mgr.14201) 194 : cluster [DBG] pgmap v130: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:58:06.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:06 vm09 bash[21220]: cluster 2026-03-07T06:58:05.160637+0000 mgr.vm07.yrfcuj (mgr.14201) 194 : cluster [DBG] pgmap v130: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:58:06.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:06 vm09 bash[21220]: audit 2026-03-07T06:58:06.182957+0000 mon.vm07 (mon.0) 773 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:06.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:06 vm09 bash[21220]: audit 2026-03-07T06:58:06.182957+0000 mon.vm07 (mon.0) 773 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:06.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:06 vm09 bash[21220]: audit 2026-03-07T06:58:06.190461+0000 mon.vm07 (mon.0) 774 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:06 vm09 bash[21220]: audit 2026-03-07T06:58:06.190461+0000 mon.vm07 (mon.0) 774 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:06.657 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:06 vm07 bash[17031]: cluster 2026-03-07T06:58:05.160637+0000 mgr.vm07.yrfcuj (mgr.14201) 194 : cluster [DBG] pgmap v130: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:58:06.657 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:06 vm07 bash[17031]: cluster 2026-03-07T06:58:05.160637+0000 mgr.vm07.yrfcuj (mgr.14201) 194 : cluster [DBG] pgmap v130: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:58:06.657 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:06 vm07 bash[17031]: audit 2026-03-07T06:58:06.182957+0000 mon.vm07 (mon.0) 773 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:06.657 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:06 vm07 bash[17031]: audit 2026-03-07T06:58:06.182957+0000 mon.vm07 (mon.0) 773 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:06.657 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:06 vm07 bash[17031]: audit 2026-03-07T06:58:06.190461+0000 mon.vm07 (mon.0) 774 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:06.657 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:06 vm07 bash[17031]: audit 2026-03-07T06:58:06.190461+0000 mon.vm07 (mon.0) 774 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.438 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:07 vm07 bash[17031]: audit 2026-03-07T06:58:06.237810+0000 mon.vm07 (mon.0) 775 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.438 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:07 vm07 bash[17031]: audit 2026-03-07T06:58:06.237810+0000 mon.vm07 (mon.0) 775 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.438 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:07 vm07 bash[17031]: audit 2026-03-07T06:58:06.242664+0000 mon.vm07 (mon.0) 776 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.438 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:07 vm07 bash[17031]: audit 2026-03-07T06:58:06.242664+0000 mon.vm07 (mon.0) 776 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.438 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:07 vm07 bash[17031]: audit 2026-03-07T06:58:06.541082+0000 mon.vm07 (mon.0) 777 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:58:07.438 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:07 vm07 bash[17031]: audit 2026-03-07T06:58:06.541082+0000 mon.vm07 (mon.0) 777 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:58:07.438 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:07 vm07 bash[17031]: audit 2026-03-07T06:58:06.541781+0000 mon.vm07 (mon.0) 778 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:58:07.438 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:07 vm07 bash[17031]: audit 2026-03-07T06:58:06.541781+0000 mon.vm07 (mon.0) 778 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:58:07.438 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:07 vm07 bash[17031]: audit 2026-03-07T06:58:06.806546+0000 mon.vm07 (mon.0) 779 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.438 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:07 vm07 bash[17031]: audit 2026-03-07T06:58:06.806546+0000 mon.vm07 (mon.0) 779 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.438 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:07 vm07 bash[17031]: audit 2026-03-07T06:58:06.812339+0000 mon.vm07 (mon.0) 780 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.438 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:07 vm07 bash[17031]: audit 2026-03-07T06:58:06.812339+0000 mon.vm07 (mon.0) 780 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.438 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:07 vm07 bash[17031]: audit 2026-03-07T06:58:06.822124+0000 mon.vm07 (mon.0) 781 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.439 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:07 vm07 bash[17031]: audit 2026-03-07T06:58:06.822124+0000 mon.vm07 (mon.0) 781 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.439 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:07 vm07 bash[17031]: audit 2026-03-07T06:58:06.824055+0000 mon.vm07 (mon.0) 782 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:58:07.439 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:07 vm07 bash[17031]: audit 2026-03-07T06:58:06.824055+0000 mon.vm07 (mon.0) 782 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:58:07.526 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:58:07.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:07 vm09 bash[21220]: audit 2026-03-07T06:58:06.237810+0000 mon.vm07 (mon.0) 775 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:07 vm09 bash[21220]: audit 2026-03-07T06:58:06.237810+0000 mon.vm07 (mon.0) 775 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:07 vm09 bash[21220]: audit 2026-03-07T06:58:06.242664+0000 mon.vm07 (mon.0) 776 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:07 vm09 bash[21220]: audit 2026-03-07T06:58:06.242664+0000 mon.vm07 (mon.0) 776 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:07 vm09 bash[21220]: audit 2026-03-07T06:58:06.541082+0000 mon.vm07 (mon.0) 777 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:58:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:07 vm09 bash[21220]: audit 2026-03-07T06:58:06.541082+0000 mon.vm07 (mon.0) 777 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:58:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:07 vm09 bash[21220]: audit 2026-03-07T06:58:06.541781+0000 mon.vm07 (mon.0) 778 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:58:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:07 vm09 bash[21220]: audit 2026-03-07T06:58:06.541781+0000 mon.vm07 (mon.0) 778 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:58:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:07 vm09 bash[21220]: audit 2026-03-07T06:58:06.806546+0000 mon.vm07 (mon.0) 779 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:07 vm09 bash[21220]: audit 2026-03-07T06:58:06.806546+0000 mon.vm07 (mon.0) 779 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:07 vm09 bash[21220]: audit 2026-03-07T06:58:06.812339+0000 mon.vm07 (mon.0) 780 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:07 vm09 bash[21220]: audit 2026-03-07T06:58:06.812339+0000 mon.vm07 (mon.0) 780 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:07 vm09 bash[21220]: audit 2026-03-07T06:58:06.822124+0000 mon.vm07 (mon.0) 781 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:07 vm09 bash[21220]: audit 2026-03-07T06:58:06.822124+0000 mon.vm07 (mon.0) 781 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:07 vm09 bash[21220]: audit 2026-03-07T06:58:06.824055+0000 mon.vm07 (mon.0) 782 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:58:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:07 vm09 bash[21220]: audit 2026-03-07T06:58:06.824055+0000 mon.vm07 (mon.0) 782 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:58:08.098 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:58:08.098 INFO:teuthology.orchestra.run.vm07.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-07T06:53:25.695727Z", "last_refresh": "2026-03-07T06:58:06.232956Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-07T06:54:32.927755Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-07T06:53:24.247577Z", "last_refresh": "2026-03-07T06:58:06.177314Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:33.727624Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-07T06:53:23.837568Z", "last_refresh": "2026-03-07T06:58:06.177372Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-07T06:53:24.987960Z", "last_refresh": "2026-03-07T06:58:06.232714Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-07T06:58:00.908906Z service:ingress.rgw.foo [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "rgw.foo", "service_name": "ingress.rgw.foo", "service_type": "ingress", "spec": {"backend_service": "rgw.foo", "first_virtual_router_id": 50, "frontend_port": 9000, "monitor_port": 9001, "virtual_ip": "12.12.1.107/22"}, "status": {"created": "2026-03-07T06:57:29.846008Z", "last_refresh": "2026-03-07T06:58:06.177566Z", "ports": [9000, 9001], "running": 4, "size": 4, "virtual_ip": "12.12.1.107/22"}}, {"events": ["2026-03-07T06:54:35.185395Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-07T06:53:23.396871Z", "last_refresh": "2026-03-07T06:58:06.177432Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:36.226821Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm07:192.168.123.107=vm07", "vm09:192.168.123.109=vm09"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-07T06:54:13.493474Z", "last_refresh": "2026-03-07T06:58:06.177343Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:34.440095Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-07T06:53:25.333423Z", "last_refresh": "2026-03-07T06:58:06.177539Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:54.121213Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-07T06:54:54.116553Z", "last_refresh": "2026-03-07T06:58:06.177227Z", "running": 8, "size": 8}}, {"events": ["2026-03-07T06:54:36.229563Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-07T06:53:24.610265Z", "ports": [9095], "running": 0, "size": 1}}, {"events": ["2026-03-07T06:57:39.780378Z service:rgw.foo [INFO] \"service was created\""], "placement": {"count": 4, "host_pattern": "*"}, "service_id": "foo", "service_name": "rgw.foo", "service_type": "rgw", "spec": {"rgw_frontend_port": 8000}, "status": {"created": "2026-03-07T06:57:39.774719Z", "last_refresh": "2026-03-07T06:58:06.177403Z", "ports": [8000], "running": 4, "size": 4}}] 2026-03-07T07:58:08.157 INFO:tasks.cephadm:rgw.foo has 4/4 2026-03-07T07:58:08.158 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-07T07:58:08.160 INFO:tasks.cephadm:Waiting for ceph service ingress.rgw.foo to start (timeout 300)... 2026-03-07T07:58:08.160 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- ceph orch ls -f json 2026-03-07T07:58:08.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:08 vm09 bash[21220]: cephadm 2026-03-07T06:58:06.544685+0000 mgr.vm07.yrfcuj (mgr.14201) 195 : cephadm [INF] Checking dashboard <-> RGW credentials 2026-03-07T07:58:08.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:08 vm09 bash[21220]: cephadm 2026-03-07T06:58:06.544685+0000 mgr.vm07.yrfcuj (mgr.14201) 195 : cephadm [INF] Checking dashboard <-> RGW credentials 2026-03-07T07:58:08.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:08 vm09 bash[21220]: cephadm 2026-03-07T06:58:06.836273+0000 mgr.vm07.yrfcuj (mgr.14201) 196 : cephadm [INF] Reconfiguring prometheus.vm07 (dependencies changed)... 2026-03-07T07:58:08.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:08 vm09 bash[21220]: cephadm 2026-03-07T06:58:06.836273+0000 mgr.vm07.yrfcuj (mgr.14201) 196 : cephadm [INF] Reconfiguring prometheus.vm07 (dependencies changed)... 2026-03-07T07:58:08.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:08 vm09 bash[21220]: cephadm 2026-03-07T06:58:06.992151+0000 mgr.vm07.yrfcuj (mgr.14201) 197 : cephadm [INF] Reconfiguring daemon prometheus.vm07 on vm07 2026-03-07T07:58:08.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:08 vm09 bash[21220]: cephadm 2026-03-07T06:58:06.992151+0000 mgr.vm07.yrfcuj (mgr.14201) 197 : cephadm [INF] Reconfiguring daemon prometheus.vm07 on vm07 2026-03-07T07:58:08.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:08 vm09 bash[21220]: cluster 2026-03-07T06:58:07.161030+0000 mgr.vm07.yrfcuj (mgr.14201) 198 : cluster [DBG] pgmap v131: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:58:08.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:08 vm09 bash[21220]: cluster 2026-03-07T06:58:07.161030+0000 mgr.vm07.yrfcuj (mgr.14201) 198 : cluster [DBG] pgmap v131: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:58:08.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:08 vm09 bash[21220]: audit 2026-03-07T06:58:07.701763+0000 mon.vm07 (mon.0) 783 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:08.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:08 vm09 bash[21220]: audit 2026-03-07T06:58:07.701763+0000 mon.vm07 (mon.0) 783 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:08.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:08 vm09 bash[21220]: audit 2026-03-07T06:58:07.707739+0000 mon.vm07 (mon.0) 784 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:08.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:08 vm09 bash[21220]: audit 2026-03-07T06:58:07.707739+0000 mon.vm07 (mon.0) 784 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:08.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:08 vm09 bash[21220]: audit 2026-03-07T06:58:07.712530+0000 mon.vm07 (mon.0) 785 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-07T07:58:08.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:08 vm09 bash[21220]: audit 2026-03-07T06:58:07.712530+0000 mon.vm07 (mon.0) 785 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-07T07:58:08.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:08 vm09 bash[21220]: audit 2026-03-07T06:58:07.757955+0000 mon.vm07 (mon.0) 786 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:58:08.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:08 vm09 bash[21220]: audit 2026-03-07T06:58:07.757955+0000 mon.vm07 (mon.0) 786 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:58:08.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:08 vm07 bash[17031]: cephadm 2026-03-07T06:58:06.544685+0000 mgr.vm07.yrfcuj (mgr.14201) 195 : cephadm [INF] Checking dashboard <-> RGW credentials 2026-03-07T07:58:08.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:08 vm07 bash[17031]: cephadm 2026-03-07T06:58:06.544685+0000 mgr.vm07.yrfcuj (mgr.14201) 195 : cephadm [INF] Checking dashboard <-> RGW credentials 2026-03-07T07:58:08.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:08 vm07 bash[17031]: cephadm 2026-03-07T06:58:06.836273+0000 mgr.vm07.yrfcuj (mgr.14201) 196 : cephadm [INF] Reconfiguring prometheus.vm07 (dependencies changed)... 2026-03-07T07:58:08.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:08 vm07 bash[17031]: cephadm 2026-03-07T06:58:06.836273+0000 mgr.vm07.yrfcuj (mgr.14201) 196 : cephadm [INF] Reconfiguring prometheus.vm07 (dependencies changed)... 2026-03-07T07:58:08.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:08 vm07 bash[17031]: cephadm 2026-03-07T06:58:06.992151+0000 mgr.vm07.yrfcuj (mgr.14201) 197 : cephadm [INF] Reconfiguring daemon prometheus.vm07 on vm07 2026-03-07T07:58:08.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:08 vm07 bash[17031]: cephadm 2026-03-07T06:58:06.992151+0000 mgr.vm07.yrfcuj (mgr.14201) 197 : cephadm [INF] Reconfiguring daemon prometheus.vm07 on vm07 2026-03-07T07:58:08.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:08 vm07 bash[17031]: cluster 2026-03-07T06:58:07.161030+0000 mgr.vm07.yrfcuj (mgr.14201) 198 : cluster [DBG] pgmap v131: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:58:08.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:08 vm07 bash[17031]: cluster 2026-03-07T06:58:07.161030+0000 mgr.vm07.yrfcuj (mgr.14201) 198 : cluster [DBG] pgmap v131: 129 pgs: 129 active+clean; 454 KiB data, 226 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:58:08.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:08 vm07 bash[17031]: audit 2026-03-07T06:58:07.701763+0000 mon.vm07 (mon.0) 783 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:08.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:08 vm07 bash[17031]: audit 2026-03-07T06:58:07.701763+0000 mon.vm07 (mon.0) 783 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:08.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:08 vm07 bash[17031]: audit 2026-03-07T06:58:07.707739+0000 mon.vm07 (mon.0) 784 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:08.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:08 vm07 bash[17031]: audit 2026-03-07T06:58:07.707739+0000 mon.vm07 (mon.0) 784 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:08.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:08 vm07 bash[17031]: audit 2026-03-07T06:58:07.712530+0000 mon.vm07 (mon.0) 785 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-07T07:58:08.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:08 vm07 bash[17031]: audit 2026-03-07T06:58:07.712530+0000 mon.vm07 (mon.0) 785 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-07T07:58:08.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:08 vm07 bash[17031]: audit 2026-03-07T06:58:07.757955+0000 mon.vm07 (mon.0) 786 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:58:08.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:08 vm07 bash[17031]: audit 2026-03-07T06:58:07.757955+0000 mon.vm07 (mon.0) 786 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:58:09.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:09 vm09 bash[21220]: audit 2026-03-07T06:58:07.712959+0000 mgr.vm07.yrfcuj (mgr.14201) 199 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-07T07:58:09.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:09 vm09 bash[21220]: audit 2026-03-07T06:58:07.712959+0000 mgr.vm07.yrfcuj (mgr.14201) 199 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-07T07:58:09.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:09 vm09 bash[21220]: audit 2026-03-07T06:58:08.090426+0000 mgr.vm07.yrfcuj (mgr.14201) 200 : audit [DBG] from='client.24383 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:58:09.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:09 vm09 bash[21220]: audit 2026-03-07T06:58:08.090426+0000 mgr.vm07.yrfcuj (mgr.14201) 200 : audit [DBG] from='client.24383 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:58:09.657 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:09 vm07 bash[17031]: audit 2026-03-07T06:58:07.712959+0000 mgr.vm07.yrfcuj (mgr.14201) 199 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-07T07:58:09.657 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:09 vm07 bash[17031]: audit 2026-03-07T06:58:07.712959+0000 mgr.vm07.yrfcuj (mgr.14201) 199 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-07T07:58:09.657 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:09 vm07 bash[17031]: audit 2026-03-07T06:58:08.090426+0000 mgr.vm07.yrfcuj (mgr.14201) 200 : audit [DBG] from='client.24383 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:58:09.657 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:09 vm07 bash[17031]: audit 2026-03-07T06:58:08.090426+0000 mgr.vm07.yrfcuj (mgr.14201) 200 : audit [DBG] from='client.24383 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:58:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:10 vm09 bash[21220]: cluster 2026-03-07T06:58:09.161438+0000 mgr.vm07.yrfcuj (mgr.14201) 201 : cluster [DBG] pgmap v132: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 85 B/s wr, 1 op/s 2026-03-07T07:58:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:10 vm09 bash[21220]: cluster 2026-03-07T06:58:09.161438+0000 mgr.vm07.yrfcuj (mgr.14201) 201 : cluster [DBG] pgmap v132: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 85 B/s wr, 1 op/s 2026-03-07T07:58:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:10 vm09 bash[21220]: audit 2026-03-07T06:58:10.127610+0000 mon.vm07 (mon.0) 787 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:58:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:10 vm09 bash[21220]: audit 2026-03-07T06:58:10.127610+0000 mon.vm07 (mon.0) 787 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:58:10.656 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:10 vm07 bash[17031]: cluster 2026-03-07T06:58:09.161438+0000 mgr.vm07.yrfcuj (mgr.14201) 201 : cluster [DBG] pgmap v132: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 85 B/s wr, 1 op/s 2026-03-07T07:58:10.656 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:10 vm07 bash[17031]: cluster 2026-03-07T06:58:09.161438+0000 mgr.vm07.yrfcuj (mgr.14201) 201 : cluster [DBG] pgmap v132: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 85 B/s wr, 1 op/s 2026-03-07T07:58:10.656 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:10 vm07 bash[17031]: audit 2026-03-07T06:58:10.127610+0000 mon.vm07 (mon.0) 787 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:58:10.656 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:10 vm07 bash[17031]: audit 2026-03-07T06:58:10.127610+0000 mon.vm07 (mon.0) 787 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:58:12.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:12 vm07 bash[17031]: cluster 2026-03-07T06:58:11.161744+0000 mgr.vm07.yrfcuj (mgr.14201) 202 : cluster [DBG] pgmap v133: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 85 B/s wr, 2 op/s 2026-03-07T07:58:12.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:12 vm07 bash[17031]: cluster 2026-03-07T06:58:11.161744+0000 mgr.vm07.yrfcuj (mgr.14201) 202 : cluster [DBG] pgmap v133: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 85 B/s wr, 2 op/s 2026-03-07T07:58:12.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:12 vm09 bash[21220]: cluster 2026-03-07T06:58:11.161744+0000 mgr.vm07.yrfcuj (mgr.14201) 202 : cluster [DBG] pgmap v133: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 85 B/s wr, 2 op/s 2026-03-07T07:58:12.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:12 vm09 bash[21220]: cluster 2026-03-07T06:58:11.161744+0000 mgr.vm07.yrfcuj (mgr.14201) 202 : cluster [DBG] pgmap v133: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 85 B/s wr, 2 op/s 2026-03-07T07:58:12.629 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:58:13.070 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-07T07:58:13.070 INFO:teuthology.orchestra.run.vm07.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-07T06:53:25.695727Z", "last_refresh": "2026-03-07T06:58:06.232956Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-07T06:54:32.927755Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-07T06:53:24.247577Z", "last_refresh": "2026-03-07T06:58:06.232929Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:33.727624Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-07T06:53:23.837568Z", "last_refresh": "2026-03-07T06:58:06.232848Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-07T06:53:24.987960Z", "last_refresh": "2026-03-07T06:58:06.232714Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-07T06:58:00.908906Z service:ingress.rgw.foo [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "rgw.foo", "service_name": "ingress.rgw.foo", "service_type": "ingress", "spec": {"backend_service": "rgw.foo", "first_virtual_router_id": 50, "frontend_port": 9000, "monitor_port": 9001, "virtual_ip": "12.12.1.107/22"}, "status": {"created": "2026-03-07T06:57:29.846008Z", "last_refresh": "2026-03-07T06:58:06.232656Z", "ports": [9000, 9001], "running": 4, "size": 4, "virtual_ip": "12.12.1.107/22"}}, {"events": ["2026-03-07T06:54:35.185395Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-07T06:53:23.396871Z", "last_refresh": "2026-03-07T06:58:06.232902Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:36.226821Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm07:192.168.123.107=vm07", "vm09:192.168.123.109=vm09"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-07T06:54:13.493474Z", "last_refresh": "2026-03-07T06:58:06.232535Z", "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:34.440095Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-07T06:53:25.333423Z", "last_refresh": "2026-03-07T06:58:06.232794Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-07T06:54:54.121213Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-07T06:54:54.116553Z", "last_refresh": "2026-03-07T06:58:06.232589Z", "running": 8, "size": 8}}, {"events": ["2026-03-07T06:54:36.229563Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-07T06:53:24.610265Z", "ports": [9095], "running": 0, "size": 1}}, {"events": ["2026-03-07T06:57:39.780378Z service:rgw.foo [INFO] \"service was created\""], "placement": {"count": 4, "host_pattern": "*"}, "service_id": "foo", "service_name": "rgw.foo", "service_type": "rgw", "spec": {"rgw_frontend_port": 8000}, "status": {"created": "2026-03-07T06:57:39.774719Z", "last_refresh": "2026-03-07T06:58:06.232821Z", "ports": [8000], "running": 4, "size": 4}}] 2026-03-07T07:58:13.150 INFO:tasks.cephadm:ingress.rgw.foo has 4/4 2026-03-07T07:58:13.150 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-07T07:58:13.153 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm07.local 2026-03-07T07:58:13.153 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- bash -c 'echo "Check while healthy..." 2026-03-07T07:58:13.153 DEBUG:teuthology.orchestra.run.vm07:> curl http://12.12.1.107:9000/ 2026-03-07T07:58:13.153 DEBUG:teuthology.orchestra.run.vm07:> 2026-03-07T07:58:13.153 DEBUG:teuthology.orchestra.run.vm07:> # stop each rgw in turn 2026-03-07T07:58:13.153 DEBUG:teuthology.orchestra.run.vm07:> echo "Check with each rgw stopped in turn..." 2026-03-07T07:58:13.153 DEBUG:teuthology.orchestra.run.vm07:> for rgw in `ceph orch ps | grep ^rgw.foo. | awk '"'"'{print $1}'"'"'`; do 2026-03-07T07:58:13.153 DEBUG:teuthology.orchestra.run.vm07:> ceph orch daemon stop $rgw 2026-03-07T07:58:13.153 DEBUG:teuthology.orchestra.run.vm07:> timeout 300 bash -c "while ! ceph orch ps | grep $rgw | grep stopped; do echo '"'"'Waiting for $rgw to stop'"'"'; ceph orch ps --daemon-type rgw; ceph health detail; sleep 5 ; done" 2026-03-07T07:58:13.153 DEBUG:teuthology.orchestra.run.vm07:> timeout 300 bash -c "while ! curl http://12.12.1.107:9000/ ; do echo '"'"'Waiting for http://12.12.1.107:9000/ to be available'"'"'; sleep 1 ; done" 2026-03-07T07:58:13.153 DEBUG:teuthology.orchestra.run.vm07:> ceph orch daemon start $rgw 2026-03-07T07:58:13.153 DEBUG:teuthology.orchestra.run.vm07:> timeout 300 bash -c "while ! ceph orch ps | grep $rgw | grep running; do echo '"'"'Waiting for $rgw to start'"'"'; ceph orch ps --daemon-type rgw; ceph health detail; sleep 5 ; done" 2026-03-07T07:58:13.153 DEBUG:teuthology.orchestra.run.vm07:> done 2026-03-07T07:58:13.154 DEBUG:teuthology.orchestra.run.vm07:> 2026-03-07T07:58:13.154 DEBUG:teuthology.orchestra.run.vm07:> # stop each haproxy in turn 2026-03-07T07:58:13.154 DEBUG:teuthology.orchestra.run.vm07:> echo "Check with each haproxy down in turn..." 2026-03-07T07:58:13.154 DEBUG:teuthology.orchestra.run.vm07:> for haproxy in `ceph orch ps | grep ^haproxy.rgw.foo. | awk '"'"'{print $1}'"'"'`; do 2026-03-07T07:58:13.154 DEBUG:teuthology.orchestra.run.vm07:> ceph orch daemon stop $haproxy 2026-03-07T07:58:13.154 DEBUG:teuthology.orchestra.run.vm07:> timeout 300 bash -c "while ! ceph orch ps | grep $haproxy | grep stopped; do echo '"'"'Waiting for $haproxy to stop'"'"'; ceph orch ps --daemon-type haproxy; ceph health detail; sleep 5 ; done" 2026-03-07T07:58:13.154 DEBUG:teuthology.orchestra.run.vm07:> timeout 300 bash -c "while ! curl http://12.12.1.107:9000/ ; do echo '"'"'Waiting for http://12.12.1.107:9000/ to be available'"'"'; sleep 1 ; done" 2026-03-07T07:58:13.154 DEBUG:teuthology.orchestra.run.vm07:> ceph orch daemon start $haproxy 2026-03-07T07:58:13.154 DEBUG:teuthology.orchestra.run.vm07:> timeout 300 bash -c "while ! ceph orch ps | grep $haproxy | grep running; do echo '"'"'Waiting for $haproxy to start'"'"'; ceph orch ps --daemon-type haproxy; ceph health detail; sleep 5 ; done" 2026-03-07T07:58:13.154 DEBUG:teuthology.orchestra.run.vm07:> done 2026-03-07T07:58:13.154 DEBUG:teuthology.orchestra.run.vm07:> 2026-03-07T07:58:13.154 DEBUG:teuthology.orchestra.run.vm07:> timeout 300 bash -c "while ! curl http://12.12.1.107:9000/ ; do echo '"'"'Waiting for http://12.12.1.107:9000/ to be available'"'"'; sleep 1 ; done"' 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: audit 2026-03-07T06:58:13.062561+0000 mon.vm07 (mon.0) 788 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: audit 2026-03-07T06:58:13.062561+0000 mon.vm07 (mon.0) 788 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: audit 2026-03-07T06:58:13.064458+0000 mgr.vm07.yrfcuj (mgr.14201) 203 : audit [DBG] from='client.24387 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: audit 2026-03-07T06:58:13.064458+0000 mgr.vm07.yrfcuj (mgr.14201) 203 : audit [DBG] from='client.24387 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: audit 2026-03-07T06:58:13.071154+0000 mon.vm07 (mon.0) 789 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: audit 2026-03-07T06:58:13.071154+0000 mon.vm07 (mon.0) 789 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: cluster 2026-03-07T06:58:13.162199+0000 mgr.vm07.yrfcuj (mgr.14201) 204 : cluster [DBG] pgmap v134: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 170 B/s wr, 2 op/s 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: cluster 2026-03-07T06:58:13.162199+0000 mgr.vm07.yrfcuj (mgr.14201) 204 : cluster [DBG] pgmap v134: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 170 B/s wr, 2 op/s 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: audit 2026-03-07T06:58:13.570332+0000 mon.vm07 (mon.0) 790 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: audit 2026-03-07T06:58:13.570332+0000 mon.vm07 (mon.0) 790 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: audit 2026-03-07T06:58:13.575012+0000 mon.vm07 (mon.0) 791 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: audit 2026-03-07T06:58:13.575012+0000 mon.vm07 (mon.0) 791 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: audit 2026-03-07T06:58:13.575800+0000 mon.vm07 (mon.0) 792 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: audit 2026-03-07T06:58:13.575800+0000 mon.vm07 (mon.0) 792 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: audit 2026-03-07T06:58:13.576272+0000 mon.vm07 (mon.0) 793 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: audit 2026-03-07T06:58:13.576272+0000 mon.vm07 (mon.0) 793 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: audit 2026-03-07T06:58:13.742799+0000 mon.vm07 (mon.0) 794 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: audit 2026-03-07T06:58:13.742799+0000 mon.vm07 (mon.0) 794 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: audit 2026-03-07T06:58:13.745149+0000 mon.vm07 (mon.0) 795 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:58:14.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:14 vm09 bash[21220]: audit 2026-03-07T06:58:13.745149+0000 mon.vm07 (mon.0) 795 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:58:14.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: audit 2026-03-07T06:58:13.062561+0000 mon.vm07 (mon.0) 788 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: audit 2026-03-07T06:58:13.062561+0000 mon.vm07 (mon.0) 788 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: audit 2026-03-07T06:58:13.064458+0000 mgr.vm07.yrfcuj (mgr.14201) 203 : audit [DBG] from='client.24387 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:58:14.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: audit 2026-03-07T06:58:13.064458+0000 mgr.vm07.yrfcuj (mgr.14201) 203 : audit [DBG] from='client.24387 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-07T07:58:14.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: audit 2026-03-07T06:58:13.071154+0000 mon.vm07 (mon.0) 789 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: audit 2026-03-07T06:58:13.071154+0000 mon.vm07 (mon.0) 789 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: cluster 2026-03-07T06:58:13.162199+0000 mgr.vm07.yrfcuj (mgr.14201) 204 : cluster [DBG] pgmap v134: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 170 B/s wr, 2 op/s 2026-03-07T07:58:14.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: cluster 2026-03-07T06:58:13.162199+0000 mgr.vm07.yrfcuj (mgr.14201) 204 : cluster [DBG] pgmap v134: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 170 B/s wr, 2 op/s 2026-03-07T07:58:14.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: audit 2026-03-07T06:58:13.570332+0000 mon.vm07 (mon.0) 790 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: audit 2026-03-07T06:58:13.570332+0000 mon.vm07 (mon.0) 790 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: audit 2026-03-07T06:58:13.575012+0000 mon.vm07 (mon.0) 791 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: audit 2026-03-07T06:58:13.575012+0000 mon.vm07 (mon.0) 791 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: audit 2026-03-07T06:58:13.575800+0000 mon.vm07 (mon.0) 792 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:58:14.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: audit 2026-03-07T06:58:13.575800+0000 mon.vm07 (mon.0) 792 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:58:14.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: audit 2026-03-07T06:58:13.576272+0000 mon.vm07 (mon.0) 793 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:58:14.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: audit 2026-03-07T06:58:13.576272+0000 mon.vm07 (mon.0) 793 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:58:14.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: audit 2026-03-07T06:58:13.742799+0000 mon.vm07 (mon.0) 794 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: audit 2026-03-07T06:58:13.742799+0000 mon.vm07 (mon.0) 794 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:14.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: audit 2026-03-07T06:58:13.745149+0000 mon.vm07 (mon.0) 795 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:58:14.407 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:14 vm07 bash[17031]: audit 2026-03-07T06:58:13.745149+0000 mon.vm07 (mon.0) 795 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:58:15.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:15 vm09 bash[21220]: cephadm 2026-03-07T06:58:13.578597+0000 mgr.vm07.yrfcuj (mgr.14201) 205 : cephadm [INF] Checking dashboard <-> RGW credentials 2026-03-07T07:58:15.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:15 vm09 bash[21220]: cephadm 2026-03-07T06:58:13.578597+0000 mgr.vm07.yrfcuj (mgr.14201) 205 : cephadm [INF] Checking dashboard <-> RGW credentials 2026-03-07T07:58:15.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:15 vm07 bash[17031]: cephadm 2026-03-07T06:58:13.578597+0000 mgr.vm07.yrfcuj (mgr.14201) 205 : cephadm [INF] Checking dashboard <-> RGW credentials 2026-03-07T07:58:15.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:15 vm07 bash[17031]: cephadm 2026-03-07T06:58:13.578597+0000 mgr.vm07.yrfcuj (mgr.14201) 205 : cephadm [INF] Checking dashboard <-> RGW credentials 2026-03-07T07:58:16.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:16 vm09 bash[21220]: cluster 2026-03-07T06:58:15.162622+0000 mgr.vm07.yrfcuj (mgr.14201) 206 : cluster [DBG] pgmap v135: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 170 B/s wr, 2 op/s 2026-03-07T07:58:16.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:16 vm09 bash[21220]: cluster 2026-03-07T06:58:15.162622+0000 mgr.vm07.yrfcuj (mgr.14201) 206 : cluster [DBG] pgmap v135: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 170 B/s wr, 2 op/s 2026-03-07T07:58:16.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:16 vm07 bash[17031]: cluster 2026-03-07T06:58:15.162622+0000 mgr.vm07.yrfcuj (mgr.14201) 206 : cluster [DBG] pgmap v135: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 170 B/s wr, 2 op/s 2026-03-07T07:58:16.406 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:16 vm07 bash[17031]: cluster 2026-03-07T06:58:15.162622+0000 mgr.vm07.yrfcuj (mgr.14201) 206 : cluster [DBG] pgmap v135: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.4 KiB/s rd, 170 B/s wr, 2 op/s 2026-03-07T07:58:18.011 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T07:58:18.122 INFO:teuthology.orchestra.run.vm07.stdout:Check while healthy... 2026-03-07T07:58:18.125 INFO:teuthology.orchestra.run.vm07.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-07T07:58:18.125 INFO:teuthology.orchestra.run.vm07.stderr: Dload Upload Total Spent Left Speed 2026-03-07T07:58:18.126 INFO:teuthology.orchestra.run.vm07.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 182k 0 --:--:-- --:--:-- --:--:-- 182k 2026-03-07T07:58:18.127 INFO:teuthology.orchestra.run.vm07.stdout:anonymousCheck with each rgw stopped in turn... 2026-03-07T07:58:18.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:18 vm09 bash[21220]: cluster 2026-03-07T06:58:17.162989+0000 mgr.vm07.yrfcuj (mgr.14201) 207 : cluster [DBG] pgmap v136: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.7 KiB/s rd, 853 B/s wr, 3 op/s 2026-03-07T07:58:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:18 vm09 bash[21220]: cluster 2026-03-07T06:58:17.162989+0000 mgr.vm07.yrfcuj (mgr.14201) 207 : cluster [DBG] pgmap v136: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.7 KiB/s rd, 853 B/s wr, 3 op/s 2026-03-07T07:58:18.405 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:18 vm07 bash[17031]: cluster 2026-03-07T06:58:17.162989+0000 mgr.vm07.yrfcuj (mgr.14201) 207 : cluster [DBG] pgmap v136: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.7 KiB/s rd, 853 B/s wr, 3 op/s 2026-03-07T07:58:18.405 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:18 vm07 bash[17031]: cluster 2026-03-07T06:58:17.162989+0000 mgr.vm07.yrfcuj (mgr.14201) 207 : cluster [DBG] pgmap v136: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.7 KiB/s rd, 853 B/s wr, 3 op/s 2026-03-07T07:58:18.566 INFO:teuthology.orchestra.run.vm07.stdout:Scheduled to stop rgw.foo.vm07.bgfaka on host 'vm07' 2026-03-07T07:58:18.759 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T07:58:18.927 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:58:18.927 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (41s) 5s ago 41s 92.8M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 4114607b8ccb 2026-03-07T07:58:18.927 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (39s) 5s ago 39s 92.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T07:58:18.927 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (41s) 5s ago 41s 90.6M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T07:58:18.927 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (40s) 5s ago 40s 90.7M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T07:58:19.110 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T07:58:19.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:19 vm09 bash[21220]: audit 2026-03-07T06:58:18.377321+0000 mgr.vm07.yrfcuj (mgr.14201) 208 : audit [DBG] from='client.14620 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:19.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:19 vm09 bash[21220]: audit 2026-03-07T06:58:18.377321+0000 mgr.vm07.yrfcuj (mgr.14201) 208 : audit [DBG] from='client.14620 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:19.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:19 vm09 bash[21220]: audit 2026-03-07T06:58:18.556517+0000 mon.vm07 (mon.0) 796 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:19.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:19 vm09 bash[21220]: audit 2026-03-07T06:58:18.556517+0000 mon.vm07 (mon.0) 796 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:19.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:19 vm09 bash[21220]: audit 2026-03-07T06:58:18.561922+0000 mon.vm07 (mon.0) 797 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:19.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:19 vm09 bash[21220]: audit 2026-03-07T06:58:18.561922+0000 mon.vm07 (mon.0) 797 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:19.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:19 vm09 bash[21220]: audit 2026-03-07T06:58:18.562854+0000 mon.vm07 (mon.0) 798 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:58:19.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:19 vm09 bash[21220]: audit 2026-03-07T06:58:18.562854+0000 mon.vm07 (mon.0) 798 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:58:19.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:19 vm09 bash[21220]: audit 2026-03-07T06:58:19.107336+0000 mon.vm07 (mon.0) 799 : audit [DBG] from='client.? 192.168.123.107:0/1169097220' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:19.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:19 vm09 bash[21220]: audit 2026-03-07T06:58:19.107336+0000 mon.vm07 (mon.0) 799 : audit [DBG] from='client.? 192.168.123.107:0/1169097220' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:19.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:19 vm07 bash[17031]: audit 2026-03-07T06:58:18.377321+0000 mgr.vm07.yrfcuj (mgr.14201) 208 : audit [DBG] from='client.14620 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:19.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:19 vm07 bash[17031]: audit 2026-03-07T06:58:18.377321+0000 mgr.vm07.yrfcuj (mgr.14201) 208 : audit [DBG] from='client.14620 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:19.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:19 vm07 bash[17031]: audit 2026-03-07T06:58:18.556517+0000 mon.vm07 (mon.0) 796 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:19.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:19 vm07 bash[17031]: audit 2026-03-07T06:58:18.556517+0000 mon.vm07 (mon.0) 796 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:19.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:19 vm07 bash[17031]: audit 2026-03-07T06:58:18.561922+0000 mon.vm07 (mon.0) 797 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:19.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:19 vm07 bash[17031]: audit 2026-03-07T06:58:18.561922+0000 mon.vm07 (mon.0) 797 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:19.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:19 vm07 bash[17031]: audit 2026-03-07T06:58:18.562854+0000 mon.vm07 (mon.0) 798 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:58:19.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:19 vm07 bash[17031]: audit 2026-03-07T06:58:18.562854+0000 mon.vm07 (mon.0) 798 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:58:19.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:19 vm07 bash[17031]: audit 2026-03-07T06:58:19.107336+0000 mon.vm07 (mon.0) 799 : audit [DBG] from='client.? 192.168.123.107:0/1169097220' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:19.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:19 vm07 bash[17031]: audit 2026-03-07T06:58:19.107336+0000 mon.vm07 (mon.0) 799 : audit [DBG] from='client.? 192.168.123.107:0/1169097220' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:20.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:20 vm09 bash[21220]: audit 2026-03-07T06:58:18.550436+0000 mgr.vm07.yrfcuj (mgr.14201) 209 : audit [DBG] from='client.14624 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "rgw.foo.vm07.bgfaka", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:20.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:20 vm09 bash[21220]: audit 2026-03-07T06:58:18.550436+0000 mgr.vm07.yrfcuj (mgr.14201) 209 : audit [DBG] from='client.14624 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "rgw.foo.vm07.bgfaka", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:20.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:20 vm09 bash[21220]: cephadm 2026-03-07T06:58:18.550861+0000 mgr.vm07.yrfcuj (mgr.14201) 210 : cephadm [INF] Schedule stop daemon rgw.foo.vm07.bgfaka 2026-03-07T07:58:20.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:20 vm09 bash[21220]: cephadm 2026-03-07T06:58:18.550861+0000 mgr.vm07.yrfcuj (mgr.14201) 210 : cephadm [INF] Schedule stop daemon rgw.foo.vm07.bgfaka 2026-03-07T07:58:20.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:20 vm09 bash[21220]: audit 2026-03-07T06:58:18.739729+0000 mgr.vm07.yrfcuj (mgr.14201) 211 : audit [DBG] from='client.14628 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:20.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:20 vm09 bash[21220]: audit 2026-03-07T06:58:18.739729+0000 mgr.vm07.yrfcuj (mgr.14201) 211 : audit [DBG] from='client.14628 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:20.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:20 vm09 bash[21220]: audit 2026-03-07T06:58:18.921248+0000 mgr.vm07.yrfcuj (mgr.14201) 212 : audit [DBG] from='client.14632 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:20.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:20 vm09 bash[21220]: audit 2026-03-07T06:58:18.921248+0000 mgr.vm07.yrfcuj (mgr.14201) 212 : audit [DBG] from='client.14632 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:20.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:20 vm09 bash[21220]: cluster 2026-03-07T06:58:19.163566+0000 mgr.vm07.yrfcuj (mgr.14201) 213 : cluster [DBG] pgmap v137: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 3.8 KiB/s rd, 853 B/s wr, 4 op/s 2026-03-07T07:58:20.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:20 vm09 bash[21220]: cluster 2026-03-07T06:58:19.163566+0000 mgr.vm07.yrfcuj (mgr.14201) 213 : cluster [DBG] pgmap v137: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 3.8 KiB/s rd, 853 B/s wr, 4 op/s 2026-03-07T07:58:20.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:20 vm07 bash[17031]: audit 2026-03-07T06:58:18.550436+0000 mgr.vm07.yrfcuj (mgr.14201) 209 : audit [DBG] from='client.14624 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "rgw.foo.vm07.bgfaka", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:20.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:20 vm07 bash[17031]: audit 2026-03-07T06:58:18.550436+0000 mgr.vm07.yrfcuj (mgr.14201) 209 : audit [DBG] from='client.14624 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "rgw.foo.vm07.bgfaka", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:20.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:20 vm07 bash[17031]: cephadm 2026-03-07T06:58:18.550861+0000 mgr.vm07.yrfcuj (mgr.14201) 210 : cephadm [INF] Schedule stop daemon rgw.foo.vm07.bgfaka 2026-03-07T07:58:20.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:20 vm07 bash[17031]: cephadm 2026-03-07T06:58:18.550861+0000 mgr.vm07.yrfcuj (mgr.14201) 210 : cephadm [INF] Schedule stop daemon rgw.foo.vm07.bgfaka 2026-03-07T07:58:20.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:20 vm07 bash[17031]: audit 2026-03-07T06:58:18.739729+0000 mgr.vm07.yrfcuj (mgr.14201) 211 : audit [DBG] from='client.14628 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:20.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:20 vm07 bash[17031]: audit 2026-03-07T06:58:18.739729+0000 mgr.vm07.yrfcuj (mgr.14201) 211 : audit [DBG] from='client.14628 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:20.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:20 vm07 bash[17031]: audit 2026-03-07T06:58:18.921248+0000 mgr.vm07.yrfcuj (mgr.14201) 212 : audit [DBG] from='client.14632 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:20.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:20 vm07 bash[17031]: audit 2026-03-07T06:58:18.921248+0000 mgr.vm07.yrfcuj (mgr.14201) 212 : audit [DBG] from='client.14632 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:20.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:20 vm07 bash[17031]: cluster 2026-03-07T06:58:19.163566+0000 mgr.vm07.yrfcuj (mgr.14201) 213 : cluster [DBG] pgmap v137: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 3.8 KiB/s rd, 853 B/s wr, 4 op/s 2026-03-07T07:58:20.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:20 vm07 bash[17031]: cluster 2026-03-07T06:58:19.163566+0000 mgr.vm07.yrfcuj (mgr.14201) 213 : cluster [DBG] pgmap v137: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 3.8 KiB/s rd, 853 B/s wr, 4 op/s 2026-03-07T07:58:22.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:22 vm09 bash[21220]: cluster 2026-03-07T06:58:21.164330+0000 mgr.vm07.yrfcuj (mgr.14201) 214 : cluster [DBG] pgmap v138: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 767 B/s wr, 2 op/s 2026-03-07T07:58:22.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:22 vm09 bash[21220]: cluster 2026-03-07T06:58:21.164330+0000 mgr.vm07.yrfcuj (mgr.14201) 214 : cluster [DBG] pgmap v138: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 767 B/s wr, 2 op/s 2026-03-07T07:58:22.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:22 vm07 bash[17031]: cluster 2026-03-07T06:58:21.164330+0000 mgr.vm07.yrfcuj (mgr.14201) 214 : cluster [DBG] pgmap v138: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 767 B/s wr, 2 op/s 2026-03-07T07:58:22.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:22 vm07 bash[17031]: cluster 2026-03-07T06:58:21.164330+0000 mgr.vm07.yrfcuj (mgr.14201) 214 : cluster [DBG] pgmap v138: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 767 B/s wr, 2 op/s 2026-03-07T07:58:24.312 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T07:58:24.480 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:58:24.480 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (46s) 0s ago 46s 93.9M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 4114607b8ccb 2026-03-07T07:58:24.480 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (44s) 0s ago 44s 93.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T07:58:24.480 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (47s) 0s ago 47s 91.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T07:58:24.480 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (45s) 0s ago 45s 91.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T07:58:24.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:24 vm09 bash[21220]: cluster 2026-03-07T06:58:23.164711+0000 mgr.vm07.yrfcuj (mgr.14201) 215 : cluster [DBG] pgmap v139: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 767 B/s wr, 2 op/s 2026-03-07T07:58:24.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:24 vm09 bash[21220]: cluster 2026-03-07T06:58:23.164711+0000 mgr.vm07.yrfcuj (mgr.14201) 215 : cluster [DBG] pgmap v139: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 767 B/s wr, 2 op/s 2026-03-07T07:58:24.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:24 vm09 bash[21220]: audit 2026-03-07T06:58:23.802709+0000 mon.vm07 (mon.0) 800 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:24 vm09 bash[21220]: audit 2026-03-07T06:58:23.802709+0000 mon.vm07 (mon.0) 800 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:24 vm09 bash[21220]: audit 2026-03-07T06:58:23.807598+0000 mon.vm07 (mon.0) 801 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:24 vm09 bash[21220]: audit 2026-03-07T06:58:23.807598+0000 mon.vm07 (mon.0) 801 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:24 vm09 bash[21220]: audit 2026-03-07T06:58:24.094645+0000 mon.vm07 (mon.0) 802 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:24 vm09 bash[21220]: audit 2026-03-07T06:58:24.094645+0000 mon.vm07 (mon.0) 802 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:24 vm09 bash[21220]: audit 2026-03-07T06:58:24.100250+0000 mon.vm07 (mon.0) 803 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:24 vm09 bash[21220]: audit 2026-03-07T06:58:24.100250+0000 mon.vm07 (mon.0) 803 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:24 vm09 bash[21220]: audit 2026-03-07T06:58:24.101282+0000 mon.vm07 (mon.0) 804 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:58:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:24 vm09 bash[21220]: audit 2026-03-07T06:58:24.101282+0000 mon.vm07 (mon.0) 804 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:58:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:24 vm09 bash[21220]: audit 2026-03-07T06:58:24.101818+0000 mon.vm07 (mon.0) 805 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:58:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:24 vm09 bash[21220]: audit 2026-03-07T06:58:24.101818+0000 mon.vm07 (mon.0) 805 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:58:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:24 vm09 bash[21220]: audit 2026-03-07T06:58:24.106045+0000 mon.vm07 (mon.0) 806 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:24 vm09 bash[21220]: audit 2026-03-07T06:58:24.106045+0000 mon.vm07 (mon.0) 806 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:24 vm09 bash[21220]: audit 2026-03-07T06:58:24.107726+0000 mon.vm07 (mon.0) 807 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:58:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:24 vm09 bash[21220]: audit 2026-03-07T06:58:24.107726+0000 mon.vm07 (mon.0) 807 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:58:24.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:24 vm07 bash[17031]: cluster 2026-03-07T06:58:23.164711+0000 mgr.vm07.yrfcuj (mgr.14201) 215 : cluster [DBG] pgmap v139: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 767 B/s wr, 2 op/s 2026-03-07T07:58:24.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:24 vm07 bash[17031]: cluster 2026-03-07T06:58:23.164711+0000 mgr.vm07.yrfcuj (mgr.14201) 215 : cluster [DBG] pgmap v139: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 767 B/s wr, 2 op/s 2026-03-07T07:58:24.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:24 vm07 bash[17031]: audit 2026-03-07T06:58:23.802709+0000 mon.vm07 (mon.0) 800 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:24 vm07 bash[17031]: audit 2026-03-07T06:58:23.802709+0000 mon.vm07 (mon.0) 800 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:24 vm07 bash[17031]: audit 2026-03-07T06:58:23.807598+0000 mon.vm07 (mon.0) 801 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:24 vm07 bash[17031]: audit 2026-03-07T06:58:23.807598+0000 mon.vm07 (mon.0) 801 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:24 vm07 bash[17031]: audit 2026-03-07T06:58:24.094645+0000 mon.vm07 (mon.0) 802 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:24 vm07 bash[17031]: audit 2026-03-07T06:58:24.094645+0000 mon.vm07 (mon.0) 802 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:24 vm07 bash[17031]: audit 2026-03-07T06:58:24.100250+0000 mon.vm07 (mon.0) 803 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:24 vm07 bash[17031]: audit 2026-03-07T06:58:24.100250+0000 mon.vm07 (mon.0) 803 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:24 vm07 bash[17031]: audit 2026-03-07T06:58:24.101282+0000 mon.vm07 (mon.0) 804 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:58:24.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:24 vm07 bash[17031]: audit 2026-03-07T06:58:24.101282+0000 mon.vm07 (mon.0) 804 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:58:24.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:24 vm07 bash[17031]: audit 2026-03-07T06:58:24.101818+0000 mon.vm07 (mon.0) 805 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:58:24.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:24 vm07 bash[17031]: audit 2026-03-07T06:58:24.101818+0000 mon.vm07 (mon.0) 805 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:58:24.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:24 vm07 bash[17031]: audit 2026-03-07T06:58:24.106045+0000 mon.vm07 (mon.0) 806 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:24 vm07 bash[17031]: audit 2026-03-07T06:58:24.106045+0000 mon.vm07 (mon.0) 806 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:24.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:24 vm07 bash[17031]: audit 2026-03-07T06:58:24.107726+0000 mon.vm07 (mon.0) 807 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:58:24.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:24 vm07 bash[17031]: audit 2026-03-07T06:58:24.107726+0000 mon.vm07 (mon.0) 807 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:58:24.672 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T07:58:25.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:25 vm09 bash[21220]: audit 2026-03-07T06:58:24.289293+0000 mgr.vm07.yrfcuj (mgr.14201) 216 : audit [DBG] from='client.14640 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:25 vm09 bash[21220]: audit 2026-03-07T06:58:24.289293+0000 mgr.vm07.yrfcuj (mgr.14201) 216 : audit [DBG] from='client.14640 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:25 vm09 bash[21220]: audit 2026-03-07T06:58:24.475176+0000 mgr.vm07.yrfcuj (mgr.14201) 217 : audit [DBG] from='client.24409 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:25 vm09 bash[21220]: audit 2026-03-07T06:58:24.475176+0000 mgr.vm07.yrfcuj (mgr.14201) 217 : audit [DBG] from='client.24409 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:25 vm09 bash[21220]: audit 2026-03-07T06:58:24.670634+0000 mon.vm07 (mon.0) 808 : audit [DBG] from='client.? 192.168.123.107:0/1792657161' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:25 vm09 bash[21220]: audit 2026-03-07T06:58:24.670634+0000 mon.vm07 (mon.0) 808 : audit [DBG] from='client.? 192.168.123.107:0/1792657161' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:25 vm09 bash[21220]: audit 2026-03-07T06:58:25.127737+0000 mon.vm07 (mon.0) 809 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:58:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:25 vm09 bash[21220]: audit 2026-03-07T06:58:25.127737+0000 mon.vm07 (mon.0) 809 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:58:25.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:25 vm07 bash[17031]: audit 2026-03-07T06:58:24.289293+0000 mgr.vm07.yrfcuj (mgr.14201) 216 : audit [DBG] from='client.14640 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:25.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:25 vm07 bash[17031]: audit 2026-03-07T06:58:24.289293+0000 mgr.vm07.yrfcuj (mgr.14201) 216 : audit [DBG] from='client.14640 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:25.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:25 vm07 bash[17031]: audit 2026-03-07T06:58:24.475176+0000 mgr.vm07.yrfcuj (mgr.14201) 217 : audit [DBG] from='client.24409 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:25.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:25 vm07 bash[17031]: audit 2026-03-07T06:58:24.475176+0000 mgr.vm07.yrfcuj (mgr.14201) 217 : audit [DBG] from='client.24409 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:25.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:25 vm07 bash[17031]: audit 2026-03-07T06:58:24.670634+0000 mon.vm07 (mon.0) 808 : audit [DBG] from='client.? 192.168.123.107:0/1792657161' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:25.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:25 vm07 bash[17031]: audit 2026-03-07T06:58:24.670634+0000 mon.vm07 (mon.0) 808 : audit [DBG] from='client.? 192.168.123.107:0/1792657161' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:25.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:25 vm07 bash[17031]: audit 2026-03-07T06:58:25.127737+0000 mon.vm07 (mon.0) 809 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:58:25.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:25 vm07 bash[17031]: audit 2026-03-07T06:58:25.127737+0000 mon.vm07 (mon.0) 809 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:58:26.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:26 vm09 bash[21220]: cluster 2026-03-07T06:58:25.164955+0000 mgr.vm07.yrfcuj (mgr.14201) 218 : cluster [DBG] pgmap v140: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 682 B/s wr, 2 op/s 2026-03-07T07:58:26.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:26 vm09 bash[21220]: cluster 2026-03-07T06:58:25.164955+0000 mgr.vm07.yrfcuj (mgr.14201) 218 : cluster [DBG] pgmap v140: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 682 B/s wr, 2 op/s 2026-03-07T07:58:26.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:26 vm07 bash[17031]: cluster 2026-03-07T06:58:25.164955+0000 mgr.vm07.yrfcuj (mgr.14201) 218 : cluster [DBG] pgmap v140: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 682 B/s wr, 2 op/s 2026-03-07T07:58:26.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:26 vm07 bash[17031]: cluster 2026-03-07T06:58:25.164955+0000 mgr.vm07.yrfcuj (mgr.14201) 218 : cluster [DBG] pgmap v140: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 682 B/s wr, 2 op/s 2026-03-07T07:58:28.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:28 vm09 bash[21220]: cluster 2026-03-07T06:58:27.165334+0000 mgr.vm07.yrfcuj (mgr.14201) 219 : cluster [DBG] pgmap v141: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.4 KiB/s rd, 682 B/s wr, 2 op/s 2026-03-07T07:58:28.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:28 vm09 bash[21220]: cluster 2026-03-07T06:58:27.165334+0000 mgr.vm07.yrfcuj (mgr.14201) 219 : cluster [DBG] pgmap v141: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.4 KiB/s rd, 682 B/s wr, 2 op/s 2026-03-07T07:58:28.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:28 vm07 bash[17031]: cluster 2026-03-07T06:58:27.165334+0000 mgr.vm07.yrfcuj (mgr.14201) 219 : cluster [DBG] pgmap v141: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.4 KiB/s rd, 682 B/s wr, 2 op/s 2026-03-07T07:58:28.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:28 vm07 bash[17031]: cluster 2026-03-07T06:58:27.165334+0000 mgr.vm07.yrfcuj (mgr.14201) 219 : cluster [DBG] pgmap v141: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.4 KiB/s rd, 682 B/s wr, 2 op/s 2026-03-07T07:58:29.847 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T07:58:29.996 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:58:29.997 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (52s) 5s ago 52s 93.9M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 4114607b8ccb 2026-03-07T07:58:29.997 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (50s) 5s ago 50s 93.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T07:58:29.997 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (52s) 6s ago 52s 91.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T07:58:29.997 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (51s) 6s ago 51s 91.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T07:58:30.189 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T07:58:30.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:30 vm09 bash[21220]: cluster 2026-03-07T06:58:29.165828+0000 mgr.vm07.yrfcuj (mgr.14201) 220 : cluster [DBG] pgmap v142: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.1 KiB/s rd, 1 op/s 2026-03-07T07:58:30.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:30 vm09 bash[21220]: cluster 2026-03-07T06:58:29.165828+0000 mgr.vm07.yrfcuj (mgr.14201) 220 : cluster [DBG] pgmap v142: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.1 KiB/s rd, 1 op/s 2026-03-07T07:58:30.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:30 vm09 bash[21220]: audit 2026-03-07T06:58:30.187848+0000 mon.vm07 (mon.0) 810 : audit [DBG] from='client.? 192.168.123.107:0/1485781601' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:30.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:30 vm09 bash[21220]: audit 2026-03-07T06:58:30.187848+0000 mon.vm07 (mon.0) 810 : audit [DBG] from='client.? 192.168.123.107:0/1485781601' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:30.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:30 vm07 bash[17031]: cluster 2026-03-07T06:58:29.165828+0000 mgr.vm07.yrfcuj (mgr.14201) 220 : cluster [DBG] pgmap v142: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.1 KiB/s rd, 1 op/s 2026-03-07T07:58:30.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:30 vm07 bash[17031]: cluster 2026-03-07T06:58:29.165828+0000 mgr.vm07.yrfcuj (mgr.14201) 220 : cluster [DBG] pgmap v142: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 1.1 KiB/s rd, 1 op/s 2026-03-07T07:58:30.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:30 vm07 bash[17031]: audit 2026-03-07T06:58:30.187848+0000 mon.vm07 (mon.0) 810 : audit [DBG] from='client.? 192.168.123.107:0/1485781601' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:30.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:30 vm07 bash[17031]: audit 2026-03-07T06:58:30.187848+0000 mon.vm07 (mon.0) 810 : audit [DBG] from='client.? 192.168.123.107:0/1485781601' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:31.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:31 vm09 bash[21220]: audit 2026-03-07T06:58:29.830585+0000 mgr.vm07.yrfcuj (mgr.14201) 221 : audit [DBG] from='client.14652 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:31.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:31 vm09 bash[21220]: audit 2026-03-07T06:58:29.830585+0000 mgr.vm07.yrfcuj (mgr.14201) 221 : audit [DBG] from='client.14652 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:31.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:31 vm09 bash[21220]: audit 2026-03-07T06:58:29.992543+0000 mgr.vm07.yrfcuj (mgr.14201) 222 : audit [DBG] from='client.14656 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:31.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:31 vm09 bash[21220]: audit 2026-03-07T06:58:29.992543+0000 mgr.vm07.yrfcuj (mgr.14201) 222 : audit [DBG] from='client.14656 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:31.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:31 vm07 bash[17031]: audit 2026-03-07T06:58:29.830585+0000 mgr.vm07.yrfcuj (mgr.14201) 221 : audit [DBG] from='client.14652 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:31.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:31 vm07 bash[17031]: audit 2026-03-07T06:58:29.830585+0000 mgr.vm07.yrfcuj (mgr.14201) 221 : audit [DBG] from='client.14652 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:31.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:31 vm07 bash[17031]: audit 2026-03-07T06:58:29.992543+0000 mgr.vm07.yrfcuj (mgr.14201) 222 : audit [DBG] from='client.14656 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:31.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:31 vm07 bash[17031]: audit 2026-03-07T06:58:29.992543+0000 mgr.vm07.yrfcuj (mgr.14201) 222 : audit [DBG] from='client.14656 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:32.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:32 vm09 bash[21220]: cluster 2026-03-07T06:58:31.166140+0000 mgr.vm07.yrfcuj (mgr.14201) 223 : cluster [DBG] pgmap v143: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:58:32.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:32 vm09 bash[21220]: cluster 2026-03-07T06:58:31.166140+0000 mgr.vm07.yrfcuj (mgr.14201) 223 : cluster [DBG] pgmap v143: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:58:32.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:32 vm07 bash[17031]: cluster 2026-03-07T06:58:31.166140+0000 mgr.vm07.yrfcuj (mgr.14201) 223 : cluster [DBG] pgmap v143: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:58:32.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:32 vm07 bash[17031]: cluster 2026-03-07T06:58:31.166140+0000 mgr.vm07.yrfcuj (mgr.14201) 223 : cluster [DBG] pgmap v143: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:58:34.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:34 vm09 bash[21220]: cluster 2026-03-07T06:58:33.166513+0000 mgr.vm07.yrfcuj (mgr.14201) 224 : cluster [DBG] pgmap v144: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:34.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:34 vm09 bash[21220]: cluster 2026-03-07T06:58:33.166513+0000 mgr.vm07.yrfcuj (mgr.14201) 224 : cluster [DBG] pgmap v144: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:34.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:34 vm07 bash[17031]: cluster 2026-03-07T06:58:33.166513+0000 mgr.vm07.yrfcuj (mgr.14201) 224 : cluster [DBG] pgmap v144: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:34.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:34 vm07 bash[17031]: cluster 2026-03-07T06:58:33.166513+0000 mgr.vm07.yrfcuj (mgr.14201) 224 : cluster [DBG] pgmap v144: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:35.369 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T07:58:35.522 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:58:35.522 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (57s) 11s ago 57s 93.9M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 4114607b8ccb 2026-03-07T07:58:35.522 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (55s) 11s ago 55s 93.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T07:58:35.522 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (58s) 11s ago 58s 91.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T07:58:35.522 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (56s) 11s ago 56s 91.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T07:58:35.705 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T07:58:36.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:36 vm09 bash[21220]: cluster 2026-03-07T06:58:35.167166+0000 mgr.vm07.yrfcuj (mgr.14201) 225 : cluster [DBG] pgmap v145: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:36.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:36 vm09 bash[21220]: cluster 2026-03-07T06:58:35.167166+0000 mgr.vm07.yrfcuj (mgr.14201) 225 : cluster [DBG] pgmap v145: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:36.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:36 vm09 bash[21220]: audit 2026-03-07T06:58:35.353361+0000 mgr.vm07.yrfcuj (mgr.14201) 226 : audit [DBG] from='client.14664 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:36.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:36 vm09 bash[21220]: audit 2026-03-07T06:58:35.353361+0000 mgr.vm07.yrfcuj (mgr.14201) 226 : audit [DBG] from='client.14664 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:36.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:36 vm09 bash[21220]: audit 2026-03-07T06:58:35.704110+0000 mon.vm07 (mon.0) 811 : audit [DBG] from='client.? 192.168.123.107:0/3447952174' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:36.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:36 vm09 bash[21220]: audit 2026-03-07T06:58:35.704110+0000 mon.vm07 (mon.0) 811 : audit [DBG] from='client.? 192.168.123.107:0/3447952174' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:36.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:36 vm07 bash[17031]: cluster 2026-03-07T06:58:35.167166+0000 mgr.vm07.yrfcuj (mgr.14201) 225 : cluster [DBG] pgmap v145: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:36.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:36 vm07 bash[17031]: cluster 2026-03-07T06:58:35.167166+0000 mgr.vm07.yrfcuj (mgr.14201) 225 : cluster [DBG] pgmap v145: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:36.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:36 vm07 bash[17031]: audit 2026-03-07T06:58:35.353361+0000 mgr.vm07.yrfcuj (mgr.14201) 226 : audit [DBG] from='client.14664 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:36.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:36 vm07 bash[17031]: audit 2026-03-07T06:58:35.353361+0000 mgr.vm07.yrfcuj (mgr.14201) 226 : audit [DBG] from='client.14664 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:36.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:36 vm07 bash[17031]: audit 2026-03-07T06:58:35.704110+0000 mon.vm07 (mon.0) 811 : audit [DBG] from='client.? 192.168.123.107:0/3447952174' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:36.654 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:36 vm07 bash[17031]: audit 2026-03-07T06:58:35.704110+0000 mon.vm07 (mon.0) 811 : audit [DBG] from='client.? 192.168.123.107:0/3447952174' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:37.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:37 vm09 bash[21220]: audit 2026-03-07T06:58:35.518305+0000 mgr.vm07.yrfcuj (mgr.14201) 227 : audit [DBG] from='client.14668 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:37.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:37 vm09 bash[21220]: audit 2026-03-07T06:58:35.518305+0000 mgr.vm07.yrfcuj (mgr.14201) 227 : audit [DBG] from='client.14668 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:37.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:37 vm07 bash[17031]: audit 2026-03-07T06:58:35.518305+0000 mgr.vm07.yrfcuj (mgr.14201) 227 : audit [DBG] from='client.14668 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:37.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:37 vm07 bash[17031]: audit 2026-03-07T06:58:35.518305+0000 mgr.vm07.yrfcuj (mgr.14201) 227 : audit [DBG] from='client.14668 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:38.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:38 vm09 bash[21220]: cluster 2026-03-07T06:58:37.167483+0000 mgr.vm07.yrfcuj (mgr.14201) 228 : cluster [DBG] pgmap v146: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:38.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:38 vm09 bash[21220]: cluster 2026-03-07T06:58:37.167483+0000 mgr.vm07.yrfcuj (mgr.14201) 228 : cluster [DBG] pgmap v146: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:38.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:38 vm07 bash[17031]: cluster 2026-03-07T06:58:37.167483+0000 mgr.vm07.yrfcuj (mgr.14201) 228 : cluster [DBG] pgmap v146: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:38.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:38 vm07 bash[17031]: cluster 2026-03-07T06:58:37.167483+0000 mgr.vm07.yrfcuj (mgr.14201) 228 : cluster [DBG] pgmap v146: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:40.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:40 vm09 bash[21220]: audit 2026-03-07T06:58:39.076086+0000 mon.vm07 (mon.0) 812 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:40.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:40 vm09 bash[21220]: audit 2026-03-07T06:58:39.076086+0000 mon.vm07 (mon.0) 812 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:40.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:40 vm09 bash[21220]: audit 2026-03-07T06:58:39.081236+0000 mon.vm07 (mon.0) 813 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:40.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:40 vm09 bash[21220]: audit 2026-03-07T06:58:39.081236+0000 mon.vm07 (mon.0) 813 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:40.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:40 vm09 bash[21220]: audit 2026-03-07T06:58:39.110339+0000 mon.vm07 (mon.0) 814 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:58:40.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:40 vm09 bash[21220]: audit 2026-03-07T06:58:39.110339+0000 mon.vm07 (mon.0) 814 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:58:40.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:40 vm09 bash[21220]: cluster 2026-03-07T06:58:39.168127+0000 mgr.vm07.yrfcuj (mgr.14201) 229 : cluster [DBG] pgmap v147: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:40.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:40 vm09 bash[21220]: cluster 2026-03-07T06:58:39.168127+0000 mgr.vm07.yrfcuj (mgr.14201) 229 : cluster [DBG] pgmap v147: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:40.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:40 vm07 bash[17031]: audit 2026-03-07T06:58:39.076086+0000 mon.vm07 (mon.0) 812 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:40.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:40 vm07 bash[17031]: audit 2026-03-07T06:58:39.076086+0000 mon.vm07 (mon.0) 812 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:40.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:40 vm07 bash[17031]: audit 2026-03-07T06:58:39.081236+0000 mon.vm07 (mon.0) 813 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:40.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:40 vm07 bash[17031]: audit 2026-03-07T06:58:39.081236+0000 mon.vm07 (mon.0) 813 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:40.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:40 vm07 bash[17031]: audit 2026-03-07T06:58:39.110339+0000 mon.vm07 (mon.0) 814 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:58:40.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:40 vm07 bash[17031]: audit 2026-03-07T06:58:39.110339+0000 mon.vm07 (mon.0) 814 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:58:40.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:40 vm07 bash[17031]: cluster 2026-03-07T06:58:39.168127+0000 mgr.vm07.yrfcuj (mgr.14201) 229 : cluster [DBG] pgmap v147: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:40.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:40 vm07 bash[17031]: cluster 2026-03-07T06:58:39.168127+0000 mgr.vm07.yrfcuj (mgr.14201) 229 : cluster [DBG] pgmap v147: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:40.885 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T07:58:41.048 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:58:41.048 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (63s) 16s ago 63s 93.9M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 4114607b8ccb 2026-03-07T07:58:41.048 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (61s) 16s ago 61s 93.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T07:58:41.048 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (63s) 17s ago 64s 91.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T07:58:41.048 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (62s) 17s ago 62s 91.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T07:58:41.242 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T07:58:41.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:41 vm09 bash[21220]: audit 2026-03-07T06:58:40.127972+0000 mon.vm07 (mon.0) 815 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:58:41.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:41 vm09 bash[21220]: audit 2026-03-07T06:58:40.127972+0000 mon.vm07 (mon.0) 815 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:58:41.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:41 vm07 bash[17031]: audit 2026-03-07T06:58:40.127972+0000 mon.vm07 (mon.0) 815 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:58:41.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:41 vm07 bash[17031]: audit 2026-03-07T06:58:40.127972+0000 mon.vm07 (mon.0) 815 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:58:42.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:42 vm09 bash[21220]: audit 2026-03-07T06:58:40.868312+0000 mgr.vm07.yrfcuj (mgr.14201) 230 : audit [DBG] from='client.14676 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:42.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:42 vm09 bash[21220]: audit 2026-03-07T06:58:40.868312+0000 mgr.vm07.yrfcuj (mgr.14201) 230 : audit [DBG] from='client.14676 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:42.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:42 vm09 bash[21220]: audit 2026-03-07T06:58:41.045102+0000 mgr.vm07.yrfcuj (mgr.14201) 231 : audit [DBG] from='client.14680 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:42.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:42 vm09 bash[21220]: audit 2026-03-07T06:58:41.045102+0000 mgr.vm07.yrfcuj (mgr.14201) 231 : audit [DBG] from='client.14680 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:42.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:42 vm09 bash[21220]: cluster 2026-03-07T06:58:41.168409+0000 mgr.vm07.yrfcuj (mgr.14201) 232 : cluster [DBG] pgmap v148: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:42.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:42 vm09 bash[21220]: cluster 2026-03-07T06:58:41.168409+0000 mgr.vm07.yrfcuj (mgr.14201) 232 : cluster [DBG] pgmap v148: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:42.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:42 vm09 bash[21220]: audit 2026-03-07T06:58:41.241237+0000 mon.vm07 (mon.0) 816 : audit [DBG] from='client.? 192.168.123.107:0/2939373859' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:42.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:42 vm09 bash[21220]: audit 2026-03-07T06:58:41.241237+0000 mon.vm07 (mon.0) 816 : audit [DBG] from='client.? 192.168.123.107:0/2939373859' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:42.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:42 vm07 bash[17031]: audit 2026-03-07T06:58:40.868312+0000 mgr.vm07.yrfcuj (mgr.14201) 230 : audit [DBG] from='client.14676 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:42.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:42 vm07 bash[17031]: audit 2026-03-07T06:58:40.868312+0000 mgr.vm07.yrfcuj (mgr.14201) 230 : audit [DBG] from='client.14676 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:42.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:42 vm07 bash[17031]: audit 2026-03-07T06:58:41.045102+0000 mgr.vm07.yrfcuj (mgr.14201) 231 : audit [DBG] from='client.14680 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:42.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:42 vm07 bash[17031]: audit 2026-03-07T06:58:41.045102+0000 mgr.vm07.yrfcuj (mgr.14201) 231 : audit [DBG] from='client.14680 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:42.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:42 vm07 bash[17031]: cluster 2026-03-07T06:58:41.168409+0000 mgr.vm07.yrfcuj (mgr.14201) 232 : cluster [DBG] pgmap v148: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:42.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:42 vm07 bash[17031]: cluster 2026-03-07T06:58:41.168409+0000 mgr.vm07.yrfcuj (mgr.14201) 232 : cluster [DBG] pgmap v148: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:42.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:42 vm07 bash[17031]: audit 2026-03-07T06:58:41.241237+0000 mon.vm07 (mon.0) 816 : audit [DBG] from='client.? 192.168.123.107:0/2939373859' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:42.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:42 vm07 bash[17031]: audit 2026-03-07T06:58:41.241237+0000 mon.vm07 (mon.0) 816 : audit [DBG] from='client.? 192.168.123.107:0/2939373859' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:44.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:44 vm09 bash[21220]: cluster 2026-03-07T06:58:43.168826+0000 mgr.vm07.yrfcuj (mgr.14201) 233 : cluster [DBG] pgmap v149: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:44.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:44 vm09 bash[21220]: cluster 2026-03-07T06:58:43.168826+0000 mgr.vm07.yrfcuj (mgr.14201) 233 : cluster [DBG] pgmap v149: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:44.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:44 vm07 bash[17031]: cluster 2026-03-07T06:58:43.168826+0000 mgr.vm07.yrfcuj (mgr.14201) 233 : cluster [DBG] pgmap v149: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:44.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:44 vm07 bash[17031]: cluster 2026-03-07T06:58:43.168826+0000 mgr.vm07.yrfcuj (mgr.14201) 233 : cluster [DBG] pgmap v149: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T07:58:45.392 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:45 vm07 bash[17031]: audit 2026-03-07T06:58:44.372488+0000 mon.vm07 (mon.0) 817 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.392 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:45 vm07 bash[17031]: audit 2026-03-07T06:58:44.372488+0000 mon.vm07 (mon.0) 817 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.392 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:45 vm07 bash[17031]: audit 2026-03-07T06:58:44.380323+0000 mon.vm07 (mon.0) 818 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.392 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:45 vm07 bash[17031]: audit 2026-03-07T06:58:44.380323+0000 mon.vm07 (mon.0) 818 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.392 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:45 vm07 bash[17031]: audit 2026-03-07T06:58:44.702497+0000 mon.vm07 (mon.0) 819 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.392 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:45 vm07 bash[17031]: audit 2026-03-07T06:58:44.702497+0000 mon.vm07 (mon.0) 819 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.392 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:45 vm07 bash[17031]: audit 2026-03-07T06:58:44.707532+0000 mon.vm07 (mon.0) 820 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:45 vm07 bash[17031]: audit 2026-03-07T06:58:44.707532+0000 mon.vm07 (mon.0) 820 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:45 vm07 bash[17031]: audit 2026-03-07T06:58:44.708280+0000 mon.vm07 (mon.0) 821 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:58:45.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:45 vm07 bash[17031]: audit 2026-03-07T06:58:44.708280+0000 mon.vm07 (mon.0) 821 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:58:45.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:45 vm07 bash[17031]: audit 2026-03-07T06:58:44.708730+0000 mon.vm07 (mon.0) 822 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:58:45.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:45 vm07 bash[17031]: audit 2026-03-07T06:58:44.708730+0000 mon.vm07 (mon.0) 822 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:58:45.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:45 vm07 bash[17031]: audit 2026-03-07T06:58:44.711944+0000 mon.vm07 (mon.0) 823 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:45 vm07 bash[17031]: audit 2026-03-07T06:58:44.711944+0000 mon.vm07 (mon.0) 823 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:45 vm07 bash[17031]: audit 2026-03-07T06:58:44.713257+0000 mon.vm07 (mon.0) 824 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:58:45.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:45 vm07 bash[17031]: audit 2026-03-07T06:58:44.713257+0000 mon.vm07 (mon.0) 824 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:58:45.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:45 vm09 bash[21220]: audit 2026-03-07T06:58:44.372488+0000 mon.vm07 (mon.0) 817 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:45 vm09 bash[21220]: audit 2026-03-07T06:58:44.372488+0000 mon.vm07 (mon.0) 817 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:45 vm09 bash[21220]: audit 2026-03-07T06:58:44.380323+0000 mon.vm07 (mon.0) 818 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:45 vm09 bash[21220]: audit 2026-03-07T06:58:44.380323+0000 mon.vm07 (mon.0) 818 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:45 vm09 bash[21220]: audit 2026-03-07T06:58:44.702497+0000 mon.vm07 (mon.0) 819 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:45 vm09 bash[21220]: audit 2026-03-07T06:58:44.702497+0000 mon.vm07 (mon.0) 819 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:45 vm09 bash[21220]: audit 2026-03-07T06:58:44.707532+0000 mon.vm07 (mon.0) 820 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:45 vm09 bash[21220]: audit 2026-03-07T06:58:44.707532+0000 mon.vm07 (mon.0) 820 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:45 vm09 bash[21220]: audit 2026-03-07T06:58:44.708280+0000 mon.vm07 (mon.0) 821 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:58:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:45 vm09 bash[21220]: audit 2026-03-07T06:58:44.708280+0000 mon.vm07 (mon.0) 821 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:58:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:45 vm09 bash[21220]: audit 2026-03-07T06:58:44.708730+0000 mon.vm07 (mon.0) 822 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:58:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:45 vm09 bash[21220]: audit 2026-03-07T06:58:44.708730+0000 mon.vm07 (mon.0) 822 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:58:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:45 vm09 bash[21220]: audit 2026-03-07T06:58:44.711944+0000 mon.vm07 (mon.0) 823 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:45 vm09 bash[21220]: audit 2026-03-07T06:58:44.711944+0000 mon.vm07 (mon.0) 823 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:45 vm09 bash[21220]: audit 2026-03-07T06:58:44.713257+0000 mon.vm07 (mon.0) 824 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:58:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:45 vm09 bash[21220]: audit 2026-03-07T06:58:44.713257+0000 mon.vm07 (mon.0) 824 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:58:46.430 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T07:58:46.582 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:58:46.582 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 1s ago 68s - - 2026-03-07T07:58:46.582 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (66s) 1s ago 66s 94.6M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T07:58:46.582 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (69s) 2s ago 69s 92.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T07:58:46.582 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (67s) 2s ago 67s 92.4M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T07:58:46.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:46 vm07 bash[17031]: cluster 2026-03-07T06:58:44.710054+0000 mgr.vm07.yrfcuj (mgr.14201) 234 : cluster [DBG] pgmap v150: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:58:46.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:46 vm07 bash[17031]: cluster 2026-03-07T06:58:44.710054+0000 mgr.vm07.yrfcuj (mgr.14201) 234 : cluster [DBG] pgmap v150: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:58:46.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:46 vm07 bash[17031]: cluster 2026-03-07T06:58:45.706791+0000 mon.vm07 (mon.0) 825 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-07T07:58:46.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:46 vm07 bash[17031]: cluster 2026-03-07T06:58:45.706791+0000 mon.vm07 (mon.0) 825 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-07T07:58:46.775 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T07:58:46.775 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T07:58:46.775 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T07:58:46.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:46 vm09 bash[21220]: cluster 2026-03-07T06:58:44.710054+0000 mgr.vm07.yrfcuj (mgr.14201) 234 : cluster [DBG] pgmap v150: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:58:46.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:46 vm09 bash[21220]: cluster 2026-03-07T06:58:44.710054+0000 mgr.vm07.yrfcuj (mgr.14201) 234 : cluster [DBG] pgmap v150: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:58:46.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:46 vm09 bash[21220]: cluster 2026-03-07T06:58:45.706791+0000 mon.vm07 (mon.0) 825 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-07T07:58:46.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:46 vm09 bash[21220]: cluster 2026-03-07T06:58:45.706791+0000 mon.vm07 (mon.0) 825 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-07T07:58:47.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:47 vm07 bash[17031]: audit 2026-03-07T06:58:46.411662+0000 mgr.vm07.yrfcuj (mgr.14201) 235 : audit [DBG] from='client.14688 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:47.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:47 vm07 bash[17031]: audit 2026-03-07T06:58:46.411662+0000 mgr.vm07.yrfcuj (mgr.14201) 235 : audit [DBG] from='client.14688 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:47.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:47 vm07 bash[17031]: audit 2026-03-07T06:58:46.774775+0000 mon.vm07 (mon.0) 826 : audit [DBG] from='client.? 192.168.123.107:0/3195553207' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:47.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:47 vm07 bash[17031]: audit 2026-03-07T06:58:46.774775+0000 mon.vm07 (mon.0) 826 : audit [DBG] from='client.? 192.168.123.107:0/3195553207' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:47.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:47 vm09 bash[21220]: audit 2026-03-07T06:58:46.411662+0000 mgr.vm07.yrfcuj (mgr.14201) 235 : audit [DBG] from='client.14688 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:47.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:47 vm09 bash[21220]: audit 2026-03-07T06:58:46.411662+0000 mgr.vm07.yrfcuj (mgr.14201) 235 : audit [DBG] from='client.14688 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:47.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:47 vm09 bash[21220]: audit 2026-03-07T06:58:46.774775+0000 mon.vm07 (mon.0) 826 : audit [DBG] from='client.? 192.168.123.107:0/3195553207' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:47.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:47 vm09 bash[21220]: audit 2026-03-07T06:58:46.774775+0000 mon.vm07 (mon.0) 826 : audit [DBG] from='client.? 192.168.123.107:0/3195553207' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:48.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:48 vm09 bash[21220]: audit 2026-03-07T06:58:46.579461+0000 mgr.vm07.yrfcuj (mgr.14201) 236 : audit [DBG] from='client.14692 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:48.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:48 vm09 bash[21220]: audit 2026-03-07T06:58:46.579461+0000 mgr.vm07.yrfcuj (mgr.14201) 236 : audit [DBG] from='client.14692 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:48.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:48 vm09 bash[21220]: cluster 2026-03-07T06:58:46.710496+0000 mgr.vm07.yrfcuj (mgr.14201) 237 : cluster [DBG] pgmap v151: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:48.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:48 vm09 bash[21220]: cluster 2026-03-07T06:58:46.710496+0000 mgr.vm07.yrfcuj (mgr.14201) 237 : cluster [DBG] pgmap v151: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:48.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:48 vm07 bash[17031]: audit 2026-03-07T06:58:46.579461+0000 mgr.vm07.yrfcuj (mgr.14201) 236 : audit [DBG] from='client.14692 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:48.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:48 vm07 bash[17031]: audit 2026-03-07T06:58:46.579461+0000 mgr.vm07.yrfcuj (mgr.14201) 236 : audit [DBG] from='client.14692 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:48.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:48 vm07 bash[17031]: cluster 2026-03-07T06:58:46.710496+0000 mgr.vm07.yrfcuj (mgr.14201) 237 : cluster [DBG] pgmap v151: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:48.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:48 vm07 bash[17031]: cluster 2026-03-07T06:58:46.710496+0000 mgr.vm07.yrfcuj (mgr.14201) 237 : cluster [DBG] pgmap v151: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:50.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:50 vm07 bash[17031]: cluster 2026-03-07T06:58:48.711000+0000 mgr.vm07.yrfcuj (mgr.14201) 238 : cluster [DBG] pgmap v152: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:50.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:50 vm07 bash[17031]: cluster 2026-03-07T06:58:48.711000+0000 mgr.vm07.yrfcuj (mgr.14201) 238 : cluster [DBG] pgmap v152: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:50.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:50 vm09 bash[21220]: cluster 2026-03-07T06:58:48.711000+0000 mgr.vm07.yrfcuj (mgr.14201) 238 : cluster [DBG] pgmap v152: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:50.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:50 vm09 bash[21220]: cluster 2026-03-07T06:58:48.711000+0000 mgr.vm07.yrfcuj (mgr.14201) 238 : cluster [DBG] pgmap v152: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:51.953 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T07:58:52.107 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:58:52.108 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 7s ago 74s - - 2026-03-07T07:58:52.108 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (72s) 7s ago 72s 94.6M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T07:58:52.108 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (75s) 7s ago 75s 92.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T07:58:52.108 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (73s) 7s ago 73s 92.4M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T07:58:52.294 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T07:58:52.294 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T07:58:52.294 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T07:58:52.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:52 vm09 bash[21220]: cluster 2026-03-07T06:58:50.711369+0000 mgr.vm07.yrfcuj (mgr.14201) 239 : cluster [DBG] pgmap v153: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:52.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:52 vm09 bash[21220]: cluster 2026-03-07T06:58:50.711369+0000 mgr.vm07.yrfcuj (mgr.14201) 239 : cluster [DBG] pgmap v153: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:52 vm09 bash[21220]: audit 2026-03-07T06:58:52.293782+0000 mon.vm07 (mon.0) 827 : audit [DBG] from='client.? 192.168.123.107:0/2831534749' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:52 vm09 bash[21220]: audit 2026-03-07T06:58:52.293782+0000 mon.vm07 (mon.0) 827 : audit [DBG] from='client.? 192.168.123.107:0/2831534749' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:52.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:52 vm07 bash[17031]: cluster 2026-03-07T06:58:50.711369+0000 mgr.vm07.yrfcuj (mgr.14201) 239 : cluster [DBG] pgmap v153: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:52.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:52 vm07 bash[17031]: cluster 2026-03-07T06:58:50.711369+0000 mgr.vm07.yrfcuj (mgr.14201) 239 : cluster [DBG] pgmap v153: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:52.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:52 vm07 bash[17031]: audit 2026-03-07T06:58:52.293782+0000 mon.vm07 (mon.0) 827 : audit [DBG] from='client.? 192.168.123.107:0/2831534749' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:52.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:52 vm07 bash[17031]: audit 2026-03-07T06:58:52.293782+0000 mon.vm07 (mon.0) 827 : audit [DBG] from='client.? 192.168.123.107:0/2831534749' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:53.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:53 vm09 bash[21220]: audit 2026-03-07T06:58:51.936537+0000 mgr.vm07.yrfcuj (mgr.14201) 240 : audit [DBG] from='client.14700 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:53.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:53 vm09 bash[21220]: audit 2026-03-07T06:58:51.936537+0000 mgr.vm07.yrfcuj (mgr.14201) 240 : audit [DBG] from='client.14700 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:53.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:53 vm09 bash[21220]: audit 2026-03-07T06:58:52.105037+0000 mgr.vm07.yrfcuj (mgr.14201) 241 : audit [DBG] from='client.14704 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:53.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:53 vm09 bash[21220]: audit 2026-03-07T06:58:52.105037+0000 mgr.vm07.yrfcuj (mgr.14201) 241 : audit [DBG] from='client.14704 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:53.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:53 vm07 bash[17031]: audit 2026-03-07T06:58:51.936537+0000 mgr.vm07.yrfcuj (mgr.14201) 240 : audit [DBG] from='client.14700 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:53.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:53 vm07 bash[17031]: audit 2026-03-07T06:58:51.936537+0000 mgr.vm07.yrfcuj (mgr.14201) 240 : audit [DBG] from='client.14700 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:53.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:53 vm07 bash[17031]: audit 2026-03-07T06:58:52.105037+0000 mgr.vm07.yrfcuj (mgr.14201) 241 : audit [DBG] from='client.14704 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:53.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:53 vm07 bash[17031]: audit 2026-03-07T06:58:52.105037+0000 mgr.vm07.yrfcuj (mgr.14201) 241 : audit [DBG] from='client.14704 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:54.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:54 vm09 bash[21220]: cluster 2026-03-07T06:58:52.711755+0000 mgr.vm07.yrfcuj (mgr.14201) 242 : cluster [DBG] pgmap v154: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:54.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:54 vm09 bash[21220]: cluster 2026-03-07T06:58:52.711755+0000 mgr.vm07.yrfcuj (mgr.14201) 242 : cluster [DBG] pgmap v154: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:54.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:54 vm07 bash[17031]: cluster 2026-03-07T06:58:52.711755+0000 mgr.vm07.yrfcuj (mgr.14201) 242 : cluster [DBG] pgmap v154: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:54.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:54 vm07 bash[17031]: cluster 2026-03-07T06:58:52.711755+0000 mgr.vm07.yrfcuj (mgr.14201) 242 : cluster [DBG] pgmap v154: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:56.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:56 vm07 bash[17031]: cluster 2026-03-07T06:58:54.712077+0000 mgr.vm07.yrfcuj (mgr.14201) 243 : cluster [DBG] pgmap v155: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:56.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:56 vm07 bash[17031]: cluster 2026-03-07T06:58:54.712077+0000 mgr.vm07.yrfcuj (mgr.14201) 243 : cluster [DBG] pgmap v155: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:56.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:56 vm07 bash[17031]: audit 2026-03-07T06:58:55.133485+0000 mon.vm07 (mon.0) 828 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:56.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:56 vm07 bash[17031]: audit 2026-03-07T06:58:55.133485+0000 mon.vm07 (mon.0) 828 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:56.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:56 vm07 bash[17031]: audit 2026-03-07T06:58:55.134445+0000 mon.vm07 (mon.0) 829 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:58:56.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:56 vm07 bash[17031]: audit 2026-03-07T06:58:55.134445+0000 mon.vm07 (mon.0) 829 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:58:56.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:56 vm09 bash[21220]: cluster 2026-03-07T06:58:54.712077+0000 mgr.vm07.yrfcuj (mgr.14201) 243 : cluster [DBG] pgmap v155: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:56 vm09 bash[21220]: cluster 2026-03-07T06:58:54.712077+0000 mgr.vm07.yrfcuj (mgr.14201) 243 : cluster [DBG] pgmap v155: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 266 B/s rd, 532 B/s wr, 0 op/s 2026-03-07T07:58:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:56 vm09 bash[21220]: audit 2026-03-07T06:58:55.133485+0000 mon.vm07 (mon.0) 828 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:56 vm09 bash[21220]: audit 2026-03-07T06:58:55.133485+0000 mon.vm07 (mon.0) 828 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:58:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:56 vm09 bash[21220]: audit 2026-03-07T06:58:55.134445+0000 mon.vm07 (mon.0) 829 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:58:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:56 vm09 bash[21220]: audit 2026-03-07T06:58:55.134445+0000 mon.vm07 (mon.0) 829 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:58:57.473 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T07:58:57.638 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:58:57.638 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 12s ago 79s - - 2026-03-07T07:58:57.638 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (77s) 12s ago 77s 94.6M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T07:58:57.638 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (80s) 13s ago 80s 92.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T07:58:57.638 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (78s) 13s ago 78s 92.4M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T07:58:57.845 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T07:58:57.845 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T07:58:57.845 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T07:58:58.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:58 vm09 bash[21220]: cluster 2026-03-07T06:58:56.712478+0000 mgr.vm07.yrfcuj (mgr.14201) 244 : cluster [DBG] pgmap v156: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:58:58.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:58 vm09 bash[21220]: cluster 2026-03-07T06:58:56.712478+0000 mgr.vm07.yrfcuj (mgr.14201) 244 : cluster [DBG] pgmap v156: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:58:58.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:58 vm09 bash[21220]: audit 2026-03-07T06:58:57.457182+0000 mgr.vm07.yrfcuj (mgr.14201) 245 : audit [DBG] from='client.14712 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:58.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:58 vm09 bash[21220]: audit 2026-03-07T06:58:57.457182+0000 mgr.vm07.yrfcuj (mgr.14201) 245 : audit [DBG] from='client.14712 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:58.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:58 vm09 bash[21220]: audit 2026-03-07T06:58:57.845715+0000 mon.vm07 (mon.0) 830 : audit [DBG] from='client.? 192.168.123.107:0/1849970422' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:58.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:58 vm09 bash[21220]: audit 2026-03-07T06:58:57.845715+0000 mon.vm07 (mon.0) 830 : audit [DBG] from='client.? 192.168.123.107:0/1849970422' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:58.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:58 vm07 bash[17031]: cluster 2026-03-07T06:58:56.712478+0000 mgr.vm07.yrfcuj (mgr.14201) 244 : cluster [DBG] pgmap v156: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:58:58.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:58 vm07 bash[17031]: cluster 2026-03-07T06:58:56.712478+0000 mgr.vm07.yrfcuj (mgr.14201) 244 : cluster [DBG] pgmap v156: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:58:58.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:58 vm07 bash[17031]: audit 2026-03-07T06:58:57.457182+0000 mgr.vm07.yrfcuj (mgr.14201) 245 : audit [DBG] from='client.14712 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:58.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:58 vm07 bash[17031]: audit 2026-03-07T06:58:57.457182+0000 mgr.vm07.yrfcuj (mgr.14201) 245 : audit [DBG] from='client.14712 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:58.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:58 vm07 bash[17031]: audit 2026-03-07T06:58:57.845715+0000 mon.vm07 (mon.0) 830 : audit [DBG] from='client.? 192.168.123.107:0/1849970422' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:58.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:58 vm07 bash[17031]: audit 2026-03-07T06:58:57.845715+0000 mon.vm07 (mon.0) 830 : audit [DBG] from='client.? 192.168.123.107:0/1849970422' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:58:59.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:59 vm07 bash[17031]: audit 2026-03-07T06:58:57.634613+0000 mgr.vm07.yrfcuj (mgr.14201) 246 : audit [DBG] from='client.14716 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:59.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:58:59 vm07 bash[17031]: audit 2026-03-07T06:58:57.634613+0000 mgr.vm07.yrfcuj (mgr.14201) 246 : audit [DBG] from='client.14716 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:59 vm09 bash[21220]: audit 2026-03-07T06:58:57.634613+0000 mgr.vm07.yrfcuj (mgr.14201) 246 : audit [DBG] from='client.14716 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:58:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:58:59 vm09 bash[21220]: audit 2026-03-07T06:58:57.634613+0000 mgr.vm07.yrfcuj (mgr.14201) 246 : audit [DBG] from='client.14716 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:00.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:00 vm07 bash[17031]: cluster 2026-03-07T06:58:58.713062+0000 mgr.vm07.yrfcuj (mgr.14201) 247 : cluster [DBG] pgmap v157: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:00.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:00 vm07 bash[17031]: cluster 2026-03-07T06:58:58.713062+0000 mgr.vm07.yrfcuj (mgr.14201) 247 : cluster [DBG] pgmap v157: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:00.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:00 vm09 bash[21220]: cluster 2026-03-07T06:58:58.713062+0000 mgr.vm07.yrfcuj (mgr.14201) 247 : cluster [DBG] pgmap v157: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:00.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:00 vm09 bash[21220]: cluster 2026-03-07T06:58:58.713062+0000 mgr.vm07.yrfcuj (mgr.14201) 247 : cluster [DBG] pgmap v157: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:02.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:02 vm09 bash[21220]: cluster 2026-03-07T06:59:00.713428+0000 mgr.vm07.yrfcuj (mgr.14201) 248 : cluster [DBG] pgmap v158: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:02.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:02 vm09 bash[21220]: cluster 2026-03-07T06:59:00.713428+0000 mgr.vm07.yrfcuj (mgr.14201) 248 : cluster [DBG] pgmap v158: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:02.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:02 vm07 bash[17031]: cluster 2026-03-07T06:59:00.713428+0000 mgr.vm07.yrfcuj (mgr.14201) 248 : cluster [DBG] pgmap v158: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:02.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:02 vm07 bash[17031]: cluster 2026-03-07T06:59:00.713428+0000 mgr.vm07.yrfcuj (mgr.14201) 248 : cluster [DBG] pgmap v158: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:03.028 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T07:59:03.194 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:59:03.194 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 18s ago 85s - - 2026-03-07T07:59:03.195 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (83s) 18s ago 83s 94.6M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T07:59:03.195 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (86s) 18s ago 86s 92.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T07:59:03.195 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (84s) 18s ago 84s 92.4M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T07:59:03.387 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T07:59:03.387 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T07:59:03.387 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T07:59:04.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:04 vm09 bash[21220]: cluster 2026-03-07T06:59:02.713862+0000 mgr.vm07.yrfcuj (mgr.14201) 249 : cluster [DBG] pgmap v159: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:04.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:04 vm09 bash[21220]: cluster 2026-03-07T06:59:02.713862+0000 mgr.vm07.yrfcuj (mgr.14201) 249 : cluster [DBG] pgmap v159: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:04.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:04 vm09 bash[21220]: audit 2026-03-07T06:59:03.011658+0000 mgr.vm07.yrfcuj (mgr.14201) 250 : audit [DBG] from='client.14724 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:04.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:04 vm09 bash[21220]: audit 2026-03-07T06:59:03.011658+0000 mgr.vm07.yrfcuj (mgr.14201) 250 : audit [DBG] from='client.14724 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:04.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:04 vm09 bash[21220]: audit 2026-03-07T06:59:03.190012+0000 mgr.vm07.yrfcuj (mgr.14201) 251 : audit [DBG] from='client.14728 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:04.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:04 vm09 bash[21220]: audit 2026-03-07T06:59:03.190012+0000 mgr.vm07.yrfcuj (mgr.14201) 251 : audit [DBG] from='client.14728 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:04.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:04 vm09 bash[21220]: audit 2026-03-07T06:59:03.387692+0000 mon.vm07 (mon.0) 831 : audit [DBG] from='client.? 192.168.123.107:0/108891080' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:04.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:04 vm09 bash[21220]: audit 2026-03-07T06:59:03.387692+0000 mon.vm07 (mon.0) 831 : audit [DBG] from='client.? 192.168.123.107:0/108891080' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:04.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:04 vm07 bash[17031]: cluster 2026-03-07T06:59:02.713862+0000 mgr.vm07.yrfcuj (mgr.14201) 249 : cluster [DBG] pgmap v159: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:04.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:04 vm07 bash[17031]: cluster 2026-03-07T06:59:02.713862+0000 mgr.vm07.yrfcuj (mgr.14201) 249 : cluster [DBG] pgmap v159: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:04.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:04 vm07 bash[17031]: audit 2026-03-07T06:59:03.011658+0000 mgr.vm07.yrfcuj (mgr.14201) 250 : audit [DBG] from='client.14724 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:04.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:04 vm07 bash[17031]: audit 2026-03-07T06:59:03.011658+0000 mgr.vm07.yrfcuj (mgr.14201) 250 : audit [DBG] from='client.14724 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:04.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:04 vm07 bash[17031]: audit 2026-03-07T06:59:03.190012+0000 mgr.vm07.yrfcuj (mgr.14201) 251 : audit [DBG] from='client.14728 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:04.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:04 vm07 bash[17031]: audit 2026-03-07T06:59:03.190012+0000 mgr.vm07.yrfcuj (mgr.14201) 251 : audit [DBG] from='client.14728 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:04.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:04 vm07 bash[17031]: audit 2026-03-07T06:59:03.387692+0000 mon.vm07 (mon.0) 831 : audit [DBG] from='client.? 192.168.123.107:0/108891080' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:04.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:04 vm07 bash[17031]: audit 2026-03-07T06:59:03.387692+0000 mon.vm07 (mon.0) 831 : audit [DBG] from='client.? 192.168.123.107:0/108891080' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:06.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:06 vm09 bash[21220]: cluster 2026-03-07T06:59:04.714346+0000 mgr.vm07.yrfcuj (mgr.14201) 252 : cluster [DBG] pgmap v160: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:06 vm09 bash[21220]: cluster 2026-03-07T06:59:04.714346+0000 mgr.vm07.yrfcuj (mgr.14201) 252 : cluster [DBG] pgmap v160: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:06.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:06 vm07 bash[17031]: cluster 2026-03-07T06:59:04.714346+0000 mgr.vm07.yrfcuj (mgr.14201) 252 : cluster [DBG] pgmap v160: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:06.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:06 vm07 bash[17031]: cluster 2026-03-07T06:59:04.714346+0000 mgr.vm07.yrfcuj (mgr.14201) 252 : cluster [DBG] pgmap v160: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:08.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:08 vm09 bash[21220]: cluster 2026-03-07T06:59:06.714811+0000 mgr.vm07.yrfcuj (mgr.14201) 253 : cluster [DBG] pgmap v161: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:08.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:08 vm09 bash[21220]: cluster 2026-03-07T06:59:06.714811+0000 mgr.vm07.yrfcuj (mgr.14201) 253 : cluster [DBG] pgmap v161: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:08.565 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T07:59:08.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:08 vm07 bash[17031]: cluster 2026-03-07T06:59:06.714811+0000 mgr.vm07.yrfcuj (mgr.14201) 253 : cluster [DBG] pgmap v161: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:08.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:08 vm07 bash[17031]: cluster 2026-03-07T06:59:06.714811+0000 mgr.vm07.yrfcuj (mgr.14201) 253 : cluster [DBG] pgmap v161: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:08.716 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:59:08.716 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 24s ago 90s - - 2026-03-07T07:59:08.716 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (88s) 24s ago 88s 94.6M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T07:59:08.716 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (91s) 24s ago 91s 92.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T07:59:08.716 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (89s) 24s ago 90s 92.4M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T07:59:08.905 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T07:59:08.905 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T07:59:08.905 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T07:59:09.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:09 vm09 bash[21220]: audit 2026-03-07T06:59:08.906271+0000 mon.vm07 (mon.0) 832 : audit [DBG] from='client.? 192.168.123.107:0/1983152691' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:09.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:09 vm09 bash[21220]: audit 2026-03-07T06:59:08.906271+0000 mon.vm07 (mon.0) 832 : audit [DBG] from='client.? 192.168.123.107:0/1983152691' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:09.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:09 vm07 bash[17031]: audit 2026-03-07T06:59:08.906271+0000 mon.vm07 (mon.0) 832 : audit [DBG] from='client.? 192.168.123.107:0/1983152691' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:09.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:09 vm07 bash[17031]: audit 2026-03-07T06:59:08.906271+0000 mon.vm07 (mon.0) 832 : audit [DBG] from='client.? 192.168.123.107:0/1983152691' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:10 vm09 bash[21220]: audit 2026-03-07T06:59:08.550395+0000 mgr.vm07.yrfcuj (mgr.14201) 254 : audit [DBG] from='client.14736 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:10 vm09 bash[21220]: audit 2026-03-07T06:59:08.550395+0000 mgr.vm07.yrfcuj (mgr.14201) 254 : audit [DBG] from='client.14736 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:10 vm09 bash[21220]: audit 2026-03-07T06:59:08.713793+0000 mgr.vm07.yrfcuj (mgr.14201) 255 : audit [DBG] from='client.14740 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:10 vm09 bash[21220]: audit 2026-03-07T06:59:08.713793+0000 mgr.vm07.yrfcuj (mgr.14201) 255 : audit [DBG] from='client.14740 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:10 vm09 bash[21220]: cluster 2026-03-07T06:59:08.715145+0000 mgr.vm07.yrfcuj (mgr.14201) 256 : cluster [DBG] pgmap v162: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:10 vm09 bash[21220]: cluster 2026-03-07T06:59:08.715145+0000 mgr.vm07.yrfcuj (mgr.14201) 256 : cluster [DBG] pgmap v162: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:10 vm09 bash[21220]: audit 2026-03-07T06:59:10.128564+0000 mon.vm07 (mon.0) 833 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:59:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:10 vm09 bash[21220]: audit 2026-03-07T06:59:10.128564+0000 mon.vm07 (mon.0) 833 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:59:10.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:10 vm07 bash[17031]: audit 2026-03-07T06:59:08.550395+0000 mgr.vm07.yrfcuj (mgr.14201) 254 : audit [DBG] from='client.14736 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:10.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:10 vm07 bash[17031]: audit 2026-03-07T06:59:08.550395+0000 mgr.vm07.yrfcuj (mgr.14201) 254 : audit [DBG] from='client.14736 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:10.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:10 vm07 bash[17031]: audit 2026-03-07T06:59:08.713793+0000 mgr.vm07.yrfcuj (mgr.14201) 255 : audit [DBG] from='client.14740 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:10.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:10 vm07 bash[17031]: audit 2026-03-07T06:59:08.713793+0000 mgr.vm07.yrfcuj (mgr.14201) 255 : audit [DBG] from='client.14740 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:10.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:10 vm07 bash[17031]: cluster 2026-03-07T06:59:08.715145+0000 mgr.vm07.yrfcuj (mgr.14201) 256 : cluster [DBG] pgmap v162: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:10.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:10 vm07 bash[17031]: cluster 2026-03-07T06:59:08.715145+0000 mgr.vm07.yrfcuj (mgr.14201) 256 : cluster [DBG] pgmap v162: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:10.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:10 vm07 bash[17031]: audit 2026-03-07T06:59:10.128564+0000 mon.vm07 (mon.0) 833 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:59:10.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:10 vm07 bash[17031]: audit 2026-03-07T06:59:10.128564+0000 mon.vm07 (mon.0) 833 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:59:12.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:12 vm09 bash[21220]: cluster 2026-03-07T06:59:10.715498+0000 mgr.vm07.yrfcuj (mgr.14201) 257 : cluster [DBG] pgmap v163: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:12.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:12 vm09 bash[21220]: cluster 2026-03-07T06:59:10.715498+0000 mgr.vm07.yrfcuj (mgr.14201) 257 : cluster [DBG] pgmap v163: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:12.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:12 vm07 bash[17031]: cluster 2026-03-07T06:59:10.715498+0000 mgr.vm07.yrfcuj (mgr.14201) 257 : cluster [DBG] pgmap v163: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:12.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:12 vm07 bash[17031]: cluster 2026-03-07T06:59:10.715498+0000 mgr.vm07.yrfcuj (mgr.14201) 257 : cluster [DBG] pgmap v163: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:14.083 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T07:59:14.244 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:59:14.244 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 29s ago 96s - - 2026-03-07T07:59:14.244 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (94s) 29s ago 94s 94.6M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T07:59:14.244 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (97s) 29s ago 97s 92.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T07:59:14.244 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (95s) 29s ago 95s 92.4M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T07:59:14.432 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T07:59:14.432 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T07:59:14.432 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T07:59:14.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:14 vm09 bash[21220]: cluster 2026-03-07T06:59:12.715975+0000 mgr.vm07.yrfcuj (mgr.14201) 258 : cluster [DBG] pgmap v164: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:14 vm09 bash[21220]: cluster 2026-03-07T06:59:12.715975+0000 mgr.vm07.yrfcuj (mgr.14201) 258 : cluster [DBG] pgmap v164: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:14.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:14 vm07 bash[17031]: cluster 2026-03-07T06:59:12.715975+0000 mgr.vm07.yrfcuj (mgr.14201) 258 : cluster [DBG] pgmap v164: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:14.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:14 vm07 bash[17031]: cluster 2026-03-07T06:59:12.715975+0000 mgr.vm07.yrfcuj (mgr.14201) 258 : cluster [DBG] pgmap v164: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:15 vm09 bash[21220]: audit 2026-03-07T06:59:14.069217+0000 mgr.vm07.yrfcuj (mgr.14201) 259 : audit [DBG] from='client.14748 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:15 vm09 bash[21220]: audit 2026-03-07T06:59:14.069217+0000 mgr.vm07.yrfcuj (mgr.14201) 259 : audit [DBG] from='client.14748 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:15 vm09 bash[21220]: audit 2026-03-07T06:59:14.243182+0000 mgr.vm07.yrfcuj (mgr.14201) 260 : audit [DBG] from='client.14752 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:15 vm09 bash[21220]: audit 2026-03-07T06:59:14.243182+0000 mgr.vm07.yrfcuj (mgr.14201) 260 : audit [DBG] from='client.14752 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:15 vm09 bash[21220]: audit 2026-03-07T06:59:14.433345+0000 mon.vm07 (mon.0) 834 : audit [DBG] from='client.? 192.168.123.107:0/3343400802' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:15 vm09 bash[21220]: audit 2026-03-07T06:59:14.433345+0000 mon.vm07 (mon.0) 834 : audit [DBG] from='client.? 192.168.123.107:0/3343400802' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:15.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:15 vm07 bash[17031]: audit 2026-03-07T06:59:14.069217+0000 mgr.vm07.yrfcuj (mgr.14201) 259 : audit [DBG] from='client.14748 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:15.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:15 vm07 bash[17031]: audit 2026-03-07T06:59:14.069217+0000 mgr.vm07.yrfcuj (mgr.14201) 259 : audit [DBG] from='client.14748 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:15.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:15 vm07 bash[17031]: audit 2026-03-07T06:59:14.243182+0000 mgr.vm07.yrfcuj (mgr.14201) 260 : audit [DBG] from='client.14752 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:15.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:15 vm07 bash[17031]: audit 2026-03-07T06:59:14.243182+0000 mgr.vm07.yrfcuj (mgr.14201) 260 : audit [DBG] from='client.14752 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:15.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:15 vm07 bash[17031]: audit 2026-03-07T06:59:14.433345+0000 mon.vm07 (mon.0) 834 : audit [DBG] from='client.? 192.168.123.107:0/3343400802' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:15.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:15 vm07 bash[17031]: audit 2026-03-07T06:59:14.433345+0000 mon.vm07 (mon.0) 834 : audit [DBG] from='client.? 192.168.123.107:0/3343400802' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:16.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:16 vm07 bash[17031]: cluster 2026-03-07T06:59:14.716415+0000 mgr.vm07.yrfcuj (mgr.14201) 261 : cluster [DBG] pgmap v165: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:16.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:16 vm07 bash[17031]: cluster 2026-03-07T06:59:14.716415+0000 mgr.vm07.yrfcuj (mgr.14201) 261 : cluster [DBG] pgmap v165: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:16.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:16 vm09 bash[21220]: cluster 2026-03-07T06:59:14.716415+0000 mgr.vm07.yrfcuj (mgr.14201) 261 : cluster [DBG] pgmap v165: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:16 vm09 bash[21220]: cluster 2026-03-07T06:59:14.716415+0000 mgr.vm07.yrfcuj (mgr.14201) 261 : cluster [DBG] pgmap v165: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:18.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:18 vm09 bash[21220]: cluster 2026-03-07T06:59:16.716906+0000 mgr.vm07.yrfcuj (mgr.14201) 262 : cluster [DBG] pgmap v166: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:18 vm09 bash[21220]: cluster 2026-03-07T06:59:16.716906+0000 mgr.vm07.yrfcuj (mgr.14201) 262 : cluster [DBG] pgmap v166: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:18.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:18 vm07 bash[17031]: cluster 2026-03-07T06:59:16.716906+0000 mgr.vm07.yrfcuj (mgr.14201) 262 : cluster [DBG] pgmap v166: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:18.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:18 vm07 bash[17031]: cluster 2026-03-07T06:59:16.716906+0000 mgr.vm07.yrfcuj (mgr.14201) 262 : cluster [DBG] pgmap v166: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:19.615 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T07:59:19.778 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:59:19.778 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 35s ago 101s - - 2026-03-07T07:59:19.778 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (100s) 35s ago 100s 94.6M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T07:59:19.778 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (102s) 35s ago 102s 92.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T07:59:19.778 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (101s) 35s ago 101s 92.4M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T07:59:19.977 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T07:59:19.977 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T07:59:19.977 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T07:59:20.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:20 vm09 bash[21220]: cluster 2026-03-07T06:59:18.717402+0000 mgr.vm07.yrfcuj (mgr.14201) 263 : cluster [DBG] pgmap v167: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:20.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:20 vm09 bash[21220]: cluster 2026-03-07T06:59:18.717402+0000 mgr.vm07.yrfcuj (mgr.14201) 263 : cluster [DBG] pgmap v167: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:20.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:20 vm09 bash[21220]: audit 2026-03-07T06:59:19.979707+0000 mon.vm07 (mon.0) 835 : audit [DBG] from='client.? 192.168.123.107:0/528260175' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:20.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:20 vm09 bash[21220]: audit 2026-03-07T06:59:19.979707+0000 mon.vm07 (mon.0) 835 : audit [DBG] from='client.? 192.168.123.107:0/528260175' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:20.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:20 vm07 bash[17031]: cluster 2026-03-07T06:59:18.717402+0000 mgr.vm07.yrfcuj (mgr.14201) 263 : cluster [DBG] pgmap v167: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:20.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:20 vm07 bash[17031]: cluster 2026-03-07T06:59:18.717402+0000 mgr.vm07.yrfcuj (mgr.14201) 263 : cluster [DBG] pgmap v167: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:20.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:20 vm07 bash[17031]: audit 2026-03-07T06:59:19.979707+0000 mon.vm07 (mon.0) 835 : audit [DBG] from='client.? 192.168.123.107:0/528260175' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:20.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:20 vm07 bash[17031]: audit 2026-03-07T06:59:19.979707+0000 mon.vm07 (mon.0) 835 : audit [DBG] from='client.? 192.168.123.107:0/528260175' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:21.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:21 vm09 bash[21220]: audit 2026-03-07T06:59:19.599763+0000 mgr.vm07.yrfcuj (mgr.14201) 264 : audit [DBG] from='client.24475 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:21.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:21 vm09 bash[21220]: audit 2026-03-07T06:59:19.599763+0000 mgr.vm07.yrfcuj (mgr.14201) 264 : audit [DBG] from='client.24475 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:21.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:21 vm09 bash[21220]: audit 2026-03-07T06:59:19.777090+0000 mgr.vm07.yrfcuj (mgr.14201) 265 : audit [DBG] from='client.14764 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:21.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:21 vm09 bash[21220]: audit 2026-03-07T06:59:19.777090+0000 mgr.vm07.yrfcuj (mgr.14201) 265 : audit [DBG] from='client.14764 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:21.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:21 vm07 bash[17031]: audit 2026-03-07T06:59:19.599763+0000 mgr.vm07.yrfcuj (mgr.14201) 264 : audit [DBG] from='client.24475 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:21.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:21 vm07 bash[17031]: audit 2026-03-07T06:59:19.599763+0000 mgr.vm07.yrfcuj (mgr.14201) 264 : audit [DBG] from='client.24475 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:21.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:21 vm07 bash[17031]: audit 2026-03-07T06:59:19.777090+0000 mgr.vm07.yrfcuj (mgr.14201) 265 : audit [DBG] from='client.14764 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:21.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:21 vm07 bash[17031]: audit 2026-03-07T06:59:19.777090+0000 mgr.vm07.yrfcuj (mgr.14201) 265 : audit [DBG] from='client.14764 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:22.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:22 vm09 bash[21220]: cluster 2026-03-07T06:59:20.717723+0000 mgr.vm07.yrfcuj (mgr.14201) 266 : cluster [DBG] pgmap v168: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:22.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:22 vm09 bash[21220]: cluster 2026-03-07T06:59:20.717723+0000 mgr.vm07.yrfcuj (mgr.14201) 266 : cluster [DBG] pgmap v168: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:22.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:22 vm07 bash[17031]: cluster 2026-03-07T06:59:20.717723+0000 mgr.vm07.yrfcuj (mgr.14201) 266 : cluster [DBG] pgmap v168: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:22.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:22 vm07 bash[17031]: cluster 2026-03-07T06:59:20.717723+0000 mgr.vm07.yrfcuj (mgr.14201) 266 : cluster [DBG] pgmap v168: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:24 vm09 bash[21220]: cluster 2026-03-07T06:59:22.718151+0000 mgr.vm07.yrfcuj (mgr.14201) 267 : cluster [DBG] pgmap v169: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:24 vm09 bash[21220]: cluster 2026-03-07T06:59:22.718151+0000 mgr.vm07.yrfcuj (mgr.14201) 267 : cluster [DBG] pgmap v169: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:24.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:24 vm07 bash[17031]: cluster 2026-03-07T06:59:22.718151+0000 mgr.vm07.yrfcuj (mgr.14201) 267 : cluster [DBG] pgmap v169: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:24.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:24 vm07 bash[17031]: cluster 2026-03-07T06:59:22.718151+0000 mgr.vm07.yrfcuj (mgr.14201) 267 : cluster [DBG] pgmap v169: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:25.169 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T07:59:25.333 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:59:25.333 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 40s ago 107s - - 2026-03-07T07:59:25.333 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (105s) 40s ago 105s 94.6M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T07:59:25.333 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (108s) 40s ago 108s 92.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T07:59:25.333 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (106s) 40s ago 106s 92.4M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T07:59:25.525 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T07:59:25.525 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T07:59:25.525 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T07:59:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:25 vm09 bash[21220]: audit 2026-03-07T06:59:25.128673+0000 mon.vm07 (mon.0) 836 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:59:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:25 vm09 bash[21220]: audit 2026-03-07T06:59:25.128673+0000 mon.vm07 (mon.0) 836 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:59:25.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:25 vm07 bash[17031]: audit 2026-03-07T06:59:25.128673+0000 mon.vm07 (mon.0) 836 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:59:25.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:25 vm07 bash[17031]: audit 2026-03-07T06:59:25.128673+0000 mon.vm07 (mon.0) 836 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:59:26.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:26 vm09 bash[21220]: cluster 2026-03-07T06:59:24.718623+0000 mgr.vm07.yrfcuj (mgr.14201) 268 : cluster [DBG] pgmap v170: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:26.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:26 vm09 bash[21220]: cluster 2026-03-07T06:59:24.718623+0000 mgr.vm07.yrfcuj (mgr.14201) 268 : cluster [DBG] pgmap v170: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:26.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:26 vm09 bash[21220]: audit 2026-03-07T06:59:25.149703+0000 mgr.vm07.yrfcuj (mgr.14201) 269 : audit [DBG] from='client.14772 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:26.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:26 vm09 bash[21220]: audit 2026-03-07T06:59:25.149703+0000 mgr.vm07.yrfcuj (mgr.14201) 269 : audit [DBG] from='client.14772 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:26.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:26 vm09 bash[21220]: audit 2026-03-07T06:59:25.332455+0000 mgr.vm07.yrfcuj (mgr.14201) 270 : audit [DBG] from='client.14776 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:26.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:26 vm09 bash[21220]: audit 2026-03-07T06:59:25.332455+0000 mgr.vm07.yrfcuj (mgr.14201) 270 : audit [DBG] from='client.14776 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:26.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:26 vm09 bash[21220]: audit 2026-03-07T06:59:25.527482+0000 mon.vm07 (mon.0) 837 : audit [DBG] from='client.? 192.168.123.107:0/2760372261' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:26.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:26 vm09 bash[21220]: audit 2026-03-07T06:59:25.527482+0000 mon.vm07 (mon.0) 837 : audit [DBG] from='client.? 192.168.123.107:0/2760372261' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:26.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:26 vm07 bash[17031]: cluster 2026-03-07T06:59:24.718623+0000 mgr.vm07.yrfcuj (mgr.14201) 268 : cluster [DBG] pgmap v170: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:26.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:26 vm07 bash[17031]: cluster 2026-03-07T06:59:24.718623+0000 mgr.vm07.yrfcuj (mgr.14201) 268 : cluster [DBG] pgmap v170: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:26.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:26 vm07 bash[17031]: audit 2026-03-07T06:59:25.149703+0000 mgr.vm07.yrfcuj (mgr.14201) 269 : audit [DBG] from='client.14772 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:26.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:26 vm07 bash[17031]: audit 2026-03-07T06:59:25.149703+0000 mgr.vm07.yrfcuj (mgr.14201) 269 : audit [DBG] from='client.14772 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:26.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:26 vm07 bash[17031]: audit 2026-03-07T06:59:25.332455+0000 mgr.vm07.yrfcuj (mgr.14201) 270 : audit [DBG] from='client.14776 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:26.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:26 vm07 bash[17031]: audit 2026-03-07T06:59:25.332455+0000 mgr.vm07.yrfcuj (mgr.14201) 270 : audit [DBG] from='client.14776 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:26.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:26 vm07 bash[17031]: audit 2026-03-07T06:59:25.527482+0000 mon.vm07 (mon.0) 837 : audit [DBG] from='client.? 192.168.123.107:0/2760372261' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:26.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:26 vm07 bash[17031]: audit 2026-03-07T06:59:25.527482+0000 mon.vm07 (mon.0) 837 : audit [DBG] from='client.? 192.168.123.107:0/2760372261' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:28.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:28 vm09 bash[21220]: cluster 2026-03-07T06:59:26.719078+0000 mgr.vm07.yrfcuj (mgr.14201) 271 : cluster [DBG] pgmap v171: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:28.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:28 vm09 bash[21220]: cluster 2026-03-07T06:59:26.719078+0000 mgr.vm07.yrfcuj (mgr.14201) 271 : cluster [DBG] pgmap v171: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:28.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:28 vm07 bash[17031]: cluster 2026-03-07T06:59:26.719078+0000 mgr.vm07.yrfcuj (mgr.14201) 271 : cluster [DBG] pgmap v171: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:28.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:28 vm07 bash[17031]: cluster 2026-03-07T06:59:26.719078+0000 mgr.vm07.yrfcuj (mgr.14201) 271 : cluster [DBG] pgmap v171: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:30.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:30 vm09 bash[21220]: cluster 2026-03-07T06:59:28.719588+0000 mgr.vm07.yrfcuj (mgr.14201) 272 : cluster [DBG] pgmap v172: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:30.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:30 vm09 bash[21220]: cluster 2026-03-07T06:59:28.719588+0000 mgr.vm07.yrfcuj (mgr.14201) 272 : cluster [DBG] pgmap v172: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:30.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:30 vm07 bash[17031]: cluster 2026-03-07T06:59:28.719588+0000 mgr.vm07.yrfcuj (mgr.14201) 272 : cluster [DBG] pgmap v172: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:30.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:30 vm07 bash[17031]: cluster 2026-03-07T06:59:28.719588+0000 mgr.vm07.yrfcuj (mgr.14201) 272 : cluster [DBG] pgmap v172: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:30.705 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T07:59:30.861 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:59:30.861 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 46s ago 112s - - 2026-03-07T07:59:30.861 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (111s) 46s ago 111s 94.6M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T07:59:30.861 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (113s) 46s ago 113s 92.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T07:59:30.861 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (112s) 46s ago 112s 92.4M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T07:59:31.056 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T07:59:31.056 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T07:59:31.056 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T07:59:31.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:31 vm09 bash[21220]: audit 2026-03-07T06:59:31.059549+0000 mon.vm07 (mon.0) 838 : audit [DBG] from='client.? 192.168.123.107:0/3747061476' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:31.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:31 vm09 bash[21220]: audit 2026-03-07T06:59:31.059549+0000 mon.vm07 (mon.0) 838 : audit [DBG] from='client.? 192.168.123.107:0/3747061476' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:31.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:31 vm07 bash[17031]: audit 2026-03-07T06:59:31.059549+0000 mon.vm07 (mon.0) 838 : audit [DBG] from='client.? 192.168.123.107:0/3747061476' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:31.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:31 vm07 bash[17031]: audit 2026-03-07T06:59:31.059549+0000 mon.vm07 (mon.0) 838 : audit [DBG] from='client.? 192.168.123.107:0/3747061476' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:32.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:32 vm09 bash[21220]: audit 2026-03-07T06:59:30.690997+0000 mgr.vm07.yrfcuj (mgr.14201) 273 : audit [DBG] from='client.24489 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:32.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:32 vm09 bash[21220]: audit 2026-03-07T06:59:30.690997+0000 mgr.vm07.yrfcuj (mgr.14201) 273 : audit [DBG] from='client.24489 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:32.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:32 vm09 bash[21220]: cluster 2026-03-07T06:59:30.720024+0000 mgr.vm07.yrfcuj (mgr.14201) 274 : cluster [DBG] pgmap v173: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:32.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:32 vm09 bash[21220]: cluster 2026-03-07T06:59:30.720024+0000 mgr.vm07.yrfcuj (mgr.14201) 274 : cluster [DBG] pgmap v173: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:32.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:32 vm09 bash[21220]: audit 2026-03-07T06:59:30.861314+0000 mgr.vm07.yrfcuj (mgr.14201) 275 : audit [DBG] from='client.14788 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:32.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:32 vm09 bash[21220]: audit 2026-03-07T06:59:30.861314+0000 mgr.vm07.yrfcuj (mgr.14201) 275 : audit [DBG] from='client.14788 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:32.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:32 vm07 bash[17031]: audit 2026-03-07T06:59:30.690997+0000 mgr.vm07.yrfcuj (mgr.14201) 273 : audit [DBG] from='client.24489 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:32.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:32 vm07 bash[17031]: audit 2026-03-07T06:59:30.690997+0000 mgr.vm07.yrfcuj (mgr.14201) 273 : audit [DBG] from='client.24489 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:32.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:32 vm07 bash[17031]: cluster 2026-03-07T06:59:30.720024+0000 mgr.vm07.yrfcuj (mgr.14201) 274 : cluster [DBG] pgmap v173: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:32.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:32 vm07 bash[17031]: cluster 2026-03-07T06:59:30.720024+0000 mgr.vm07.yrfcuj (mgr.14201) 274 : cluster [DBG] pgmap v173: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:32.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:32 vm07 bash[17031]: audit 2026-03-07T06:59:30.861314+0000 mgr.vm07.yrfcuj (mgr.14201) 275 : audit [DBG] from='client.14788 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:32.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:32 vm07 bash[17031]: audit 2026-03-07T06:59:30.861314+0000 mgr.vm07.yrfcuj (mgr.14201) 275 : audit [DBG] from='client.14788 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:34 vm09 bash[21220]: cluster 2026-03-07T06:59:32.720562+0000 mgr.vm07.yrfcuj (mgr.14201) 276 : cluster [DBG] pgmap v174: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:34 vm09 bash[21220]: cluster 2026-03-07T06:59:32.720562+0000 mgr.vm07.yrfcuj (mgr.14201) 276 : cluster [DBG] pgmap v174: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:34.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:34 vm07 bash[17031]: cluster 2026-03-07T06:59:32.720562+0000 mgr.vm07.yrfcuj (mgr.14201) 276 : cluster [DBG] pgmap v174: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:34.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:34 vm07 bash[17031]: cluster 2026-03-07T06:59:32.720562+0000 mgr.vm07.yrfcuj (mgr.14201) 276 : cluster [DBG] pgmap v174: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:36.236 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T07:59:36.390 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:59:36.390 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 51s ago 118s - - 2026-03-07T07:59:36.390 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (116s) 51s ago 116s 94.6M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T07:59:36.390 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (119s) 52s ago 119s 92.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T07:59:36.390 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (117s) 52s ago 117s 92.4M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T07:59:36.585 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T07:59:36.585 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T07:59:36.585 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T07:59:36.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:36 vm09 bash[21220]: cluster 2026-03-07T06:59:34.721050+0000 mgr.vm07.yrfcuj (mgr.14201) 277 : cluster [DBG] pgmap v175: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:36.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:36 vm09 bash[21220]: cluster 2026-03-07T06:59:34.721050+0000 mgr.vm07.yrfcuj (mgr.14201) 277 : cluster [DBG] pgmap v175: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:36.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:36 vm07 bash[17031]: cluster 2026-03-07T06:59:34.721050+0000 mgr.vm07.yrfcuj (mgr.14201) 277 : cluster [DBG] pgmap v175: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:36.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:36 vm07 bash[17031]: cluster 2026-03-07T06:59:34.721050+0000 mgr.vm07.yrfcuj (mgr.14201) 277 : cluster [DBG] pgmap v175: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:37.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:37 vm09 bash[21220]: audit 2026-03-07T06:59:36.221509+0000 mgr.vm07.yrfcuj (mgr.14201) 278 : audit [DBG] from='client.14796 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:37.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:37 vm09 bash[21220]: audit 2026-03-07T06:59:36.221509+0000 mgr.vm07.yrfcuj (mgr.14201) 278 : audit [DBG] from='client.14796 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:37.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:37 vm09 bash[21220]: audit 2026-03-07T06:59:36.390716+0000 mgr.vm07.yrfcuj (mgr.14201) 279 : audit [DBG] from='client.14800 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:37.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:37 vm09 bash[21220]: audit 2026-03-07T06:59:36.390716+0000 mgr.vm07.yrfcuj (mgr.14201) 279 : audit [DBG] from='client.14800 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:37.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:37 vm09 bash[21220]: audit 2026-03-07T06:59:36.584980+0000 mon.vm09 (mon.1) 30 : audit [DBG] from='client.? 192.168.123.107:0/169779869' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:37.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:37 vm09 bash[21220]: audit 2026-03-07T06:59:36.584980+0000 mon.vm09 (mon.1) 30 : audit [DBG] from='client.? 192.168.123.107:0/169779869' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:37.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:37 vm07 bash[17031]: audit 2026-03-07T06:59:36.221509+0000 mgr.vm07.yrfcuj (mgr.14201) 278 : audit [DBG] from='client.14796 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:37.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:37 vm07 bash[17031]: audit 2026-03-07T06:59:36.221509+0000 mgr.vm07.yrfcuj (mgr.14201) 278 : audit [DBG] from='client.14796 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:37.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:37 vm07 bash[17031]: audit 2026-03-07T06:59:36.390716+0000 mgr.vm07.yrfcuj (mgr.14201) 279 : audit [DBG] from='client.14800 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:37.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:37 vm07 bash[17031]: audit 2026-03-07T06:59:36.390716+0000 mgr.vm07.yrfcuj (mgr.14201) 279 : audit [DBG] from='client.14800 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:37.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:37 vm07 bash[17031]: audit 2026-03-07T06:59:36.584980+0000 mon.vm09 (mon.1) 30 : audit [DBG] from='client.? 192.168.123.107:0/169779869' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:37.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:37 vm07 bash[17031]: audit 2026-03-07T06:59:36.584980+0000 mon.vm09 (mon.1) 30 : audit [DBG] from='client.? 192.168.123.107:0/169779869' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:38.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:38 vm09 bash[21220]: cluster 2026-03-07T06:59:36.721913+0000 mgr.vm07.yrfcuj (mgr.14201) 280 : cluster [DBG] pgmap v176: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:38.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:38 vm09 bash[21220]: cluster 2026-03-07T06:59:36.721913+0000 mgr.vm07.yrfcuj (mgr.14201) 280 : cluster [DBG] pgmap v176: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:38.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:38 vm07 bash[17031]: cluster 2026-03-07T06:59:36.721913+0000 mgr.vm07.yrfcuj (mgr.14201) 280 : cluster [DBG] pgmap v176: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:38.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:38 vm07 bash[17031]: cluster 2026-03-07T06:59:36.721913+0000 mgr.vm07.yrfcuj (mgr.14201) 280 : cluster [DBG] pgmap v176: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:40.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:40 vm09 bash[21220]: cluster 2026-03-07T06:59:38.722356+0000 mgr.vm07.yrfcuj (mgr.14201) 281 : cluster [DBG] pgmap v177: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:40.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:40 vm09 bash[21220]: cluster 2026-03-07T06:59:38.722356+0000 mgr.vm07.yrfcuj (mgr.14201) 281 : cluster [DBG] pgmap v177: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:40.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:40 vm09 bash[21220]: audit 2026-03-07T06:59:40.129020+0000 mon.vm07 (mon.0) 839 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:59:40.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:40 vm09 bash[21220]: audit 2026-03-07T06:59:40.129020+0000 mon.vm07 (mon.0) 839 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:59:40.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:40 vm07 bash[17031]: cluster 2026-03-07T06:59:38.722356+0000 mgr.vm07.yrfcuj (mgr.14201) 281 : cluster [DBG] pgmap v177: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:40.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:40 vm07 bash[17031]: cluster 2026-03-07T06:59:38.722356+0000 mgr.vm07.yrfcuj (mgr.14201) 281 : cluster [DBG] pgmap v177: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:40.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:40 vm07 bash[17031]: audit 2026-03-07T06:59:40.129020+0000 mon.vm07 (mon.0) 839 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:59:40.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:40 vm07 bash[17031]: audit 2026-03-07T06:59:40.129020+0000 mon.vm07 (mon.0) 839 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:59:41.767 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T07:59:41.922 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:59:41.922 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 57s ago 2m - - 2026-03-07T07:59:41.922 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 57s ago 2m 94.6M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T07:59:41.922 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 57s ago 2m 92.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T07:59:41.922 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (2m) 57s ago 2m 92.4M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T07:59:42.115 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T07:59:42.115 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T07:59:42.115 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T07:59:42.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:42 vm07 bash[17031]: cluster 2026-03-07T06:59:40.722746+0000 mgr.vm07.yrfcuj (mgr.14201) 282 : cluster [DBG] pgmap v178: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:42.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:42 vm07 bash[17031]: cluster 2026-03-07T06:59:40.722746+0000 mgr.vm07.yrfcuj (mgr.14201) 282 : cluster [DBG] pgmap v178: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:42.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:42 vm07 bash[17031]: audit 2026-03-07T06:59:42.119160+0000 mon.vm07 (mon.0) 840 : audit [DBG] from='client.? 192.168.123.107:0/3486545550' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:42.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:42 vm07 bash[17031]: audit 2026-03-07T06:59:42.119160+0000 mon.vm07 (mon.0) 840 : audit [DBG] from='client.? 192.168.123.107:0/3486545550' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:42.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:42 vm09 bash[21220]: cluster 2026-03-07T06:59:40.722746+0000 mgr.vm07.yrfcuj (mgr.14201) 282 : cluster [DBG] pgmap v178: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:42.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:42 vm09 bash[21220]: cluster 2026-03-07T06:59:40.722746+0000 mgr.vm07.yrfcuj (mgr.14201) 282 : cluster [DBG] pgmap v178: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:42.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:42 vm09 bash[21220]: audit 2026-03-07T06:59:42.119160+0000 mon.vm07 (mon.0) 840 : audit [DBG] from='client.? 192.168.123.107:0/3486545550' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:42.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:42 vm09 bash[21220]: audit 2026-03-07T06:59:42.119160+0000 mon.vm07 (mon.0) 840 : audit [DBG] from='client.? 192.168.123.107:0/3486545550' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:43.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:43 vm07 bash[17031]: audit 2026-03-07T06:59:41.754832+0000 mgr.vm07.yrfcuj (mgr.14201) 283 : audit [DBG] from='client.14808 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:43.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:43 vm07 bash[17031]: audit 2026-03-07T06:59:41.754832+0000 mgr.vm07.yrfcuj (mgr.14201) 283 : audit [DBG] from='client.14808 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:43.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:43 vm07 bash[17031]: audit 2026-03-07T06:59:41.923615+0000 mgr.vm07.yrfcuj (mgr.14201) 284 : audit [DBG] from='client.14812 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:43.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:43 vm07 bash[17031]: audit 2026-03-07T06:59:41.923615+0000 mgr.vm07.yrfcuj (mgr.14201) 284 : audit [DBG] from='client.14812 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:43.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:43 vm09 bash[21220]: audit 2026-03-07T06:59:41.754832+0000 mgr.vm07.yrfcuj (mgr.14201) 283 : audit [DBG] from='client.14808 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:43.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:43 vm09 bash[21220]: audit 2026-03-07T06:59:41.754832+0000 mgr.vm07.yrfcuj (mgr.14201) 283 : audit [DBG] from='client.14808 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:43.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:43 vm09 bash[21220]: audit 2026-03-07T06:59:41.923615+0000 mgr.vm07.yrfcuj (mgr.14201) 284 : audit [DBG] from='client.14812 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:43.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:43 vm09 bash[21220]: audit 2026-03-07T06:59:41.923615+0000 mgr.vm07.yrfcuj (mgr.14201) 284 : audit [DBG] from='client.14812 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:44.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:44 vm07 bash[17031]: cluster 2026-03-07T06:59:42.723214+0000 mgr.vm07.yrfcuj (mgr.14201) 285 : cluster [DBG] pgmap v179: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:44.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:44 vm07 bash[17031]: cluster 2026-03-07T06:59:42.723214+0000 mgr.vm07.yrfcuj (mgr.14201) 285 : cluster [DBG] pgmap v179: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:44.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:44 vm09 bash[21220]: cluster 2026-03-07T06:59:42.723214+0000 mgr.vm07.yrfcuj (mgr.14201) 285 : cluster [DBG] pgmap v179: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:44.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:44 vm09 bash[21220]: cluster 2026-03-07T06:59:42.723214+0000 mgr.vm07.yrfcuj (mgr.14201) 285 : cluster [DBG] pgmap v179: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:45.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:45 vm07 bash[17031]: audit 2026-03-07T06:59:44.755153+0000 mon.vm07 (mon.0) 841 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:59:45.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:45 vm07 bash[17031]: audit 2026-03-07T06:59:44.755153+0000 mon.vm07 (mon.0) 841 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:59:45.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:45 vm09 bash[21220]: audit 2026-03-07T06:59:44.755153+0000 mon.vm07 (mon.0) 841 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:59:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:45 vm09 bash[21220]: audit 2026-03-07T06:59:44.755153+0000 mon.vm07 (mon.0) 841 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T07:59:46.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:46 vm07 bash[17031]: cluster 2026-03-07T06:59:44.723643+0000 mgr.vm07.yrfcuj (mgr.14201) 286 : cluster [DBG] pgmap v180: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:46.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:46 vm07 bash[17031]: cluster 2026-03-07T06:59:44.723643+0000 mgr.vm07.yrfcuj (mgr.14201) 286 : cluster [DBG] pgmap v180: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:46.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:46 vm09 bash[21220]: cluster 2026-03-07T06:59:44.723643+0000 mgr.vm07.yrfcuj (mgr.14201) 286 : cluster [DBG] pgmap v180: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:46.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:46 vm09 bash[21220]: cluster 2026-03-07T06:59:44.723643+0000 mgr.vm07.yrfcuj (mgr.14201) 286 : cluster [DBG] pgmap v180: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:47.298 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T07:59:47.454 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:59:47.454 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 62s ago 2m - - 2026-03-07T07:59:47.454 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 62s ago 2m 94.6M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T07:59:47.454 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 63s ago 2m 92.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T07:59:47.454 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (2m) 63s ago 2m 92.4M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T07:59:47.649 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T07:59:47.649 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T07:59:47.649 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T07:59:48.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:48 vm07 bash[17031]: cluster 2026-03-07T06:59:46.723979+0000 mgr.vm07.yrfcuj (mgr.14201) 287 : cluster [DBG] pgmap v181: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:48.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:48 vm07 bash[17031]: cluster 2026-03-07T06:59:46.723979+0000 mgr.vm07.yrfcuj (mgr.14201) 287 : cluster [DBG] pgmap v181: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:48.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:48 vm07 bash[17031]: audit 2026-03-07T06:59:47.287702+0000 mgr.vm07.yrfcuj (mgr.14201) 288 : audit [DBG] from='client.14820 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:48.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:48 vm07 bash[17031]: audit 2026-03-07T06:59:47.287702+0000 mgr.vm07.yrfcuj (mgr.14201) 288 : audit [DBG] from='client.14820 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:48.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:48 vm07 bash[17031]: audit 2026-03-07T06:59:47.455817+0000 mgr.vm07.yrfcuj (mgr.14201) 289 : audit [DBG] from='client.14824 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:48.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:48 vm07 bash[17031]: audit 2026-03-07T06:59:47.455817+0000 mgr.vm07.yrfcuj (mgr.14201) 289 : audit [DBG] from='client.14824 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:48.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:48 vm07 bash[17031]: audit 2026-03-07T06:59:47.654143+0000 mon.vm07 (mon.0) 842 : audit [DBG] from='client.? 192.168.123.107:0/3863301071' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:48.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:48 vm07 bash[17031]: audit 2026-03-07T06:59:47.654143+0000 mon.vm07 (mon.0) 842 : audit [DBG] from='client.? 192.168.123.107:0/3863301071' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:48.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:48 vm09 bash[21220]: cluster 2026-03-07T06:59:46.723979+0000 mgr.vm07.yrfcuj (mgr.14201) 287 : cluster [DBG] pgmap v181: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:48.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:48 vm09 bash[21220]: cluster 2026-03-07T06:59:46.723979+0000 mgr.vm07.yrfcuj (mgr.14201) 287 : cluster [DBG] pgmap v181: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:48 vm09 bash[21220]: audit 2026-03-07T06:59:47.287702+0000 mgr.vm07.yrfcuj (mgr.14201) 288 : audit [DBG] from='client.14820 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:48 vm09 bash[21220]: audit 2026-03-07T06:59:47.287702+0000 mgr.vm07.yrfcuj (mgr.14201) 288 : audit [DBG] from='client.14820 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:48 vm09 bash[21220]: audit 2026-03-07T06:59:47.455817+0000 mgr.vm07.yrfcuj (mgr.14201) 289 : audit [DBG] from='client.14824 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:48 vm09 bash[21220]: audit 2026-03-07T06:59:47.455817+0000 mgr.vm07.yrfcuj (mgr.14201) 289 : audit [DBG] from='client.14824 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:48 vm09 bash[21220]: audit 2026-03-07T06:59:47.654143+0000 mon.vm07 (mon.0) 842 : audit [DBG] from='client.? 192.168.123.107:0/3863301071' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:48 vm09 bash[21220]: audit 2026-03-07T06:59:47.654143+0000 mon.vm07 (mon.0) 842 : audit [DBG] from='client.? 192.168.123.107:0/3863301071' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:50.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:50 vm07 bash[17031]: cluster 2026-03-07T06:59:48.724456+0000 mgr.vm07.yrfcuj (mgr.14201) 290 : cluster [DBG] pgmap v182: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:50.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:50 vm07 bash[17031]: cluster 2026-03-07T06:59:48.724456+0000 mgr.vm07.yrfcuj (mgr.14201) 290 : cluster [DBG] pgmap v182: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:50.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:50 vm07 bash[17031]: audit 2026-03-07T06:59:50.012927+0000 mon.vm07 (mon.0) 843 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:50.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:50 vm07 bash[17031]: audit 2026-03-07T06:59:50.012927+0000 mon.vm07 (mon.0) 843 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:50.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:50 vm07 bash[17031]: audit 2026-03-07T06:59:50.019498+0000 mon.vm07 (mon.0) 844 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:50.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:50 vm07 bash[17031]: audit 2026-03-07T06:59:50.019498+0000 mon.vm07 (mon.0) 844 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:50.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:50 vm07 bash[17031]: audit 2026-03-07T06:59:50.366273+0000 mon.vm07 (mon.0) 845 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:50.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:50 vm07 bash[17031]: audit 2026-03-07T06:59:50.366273+0000 mon.vm07 (mon.0) 845 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:50.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:50 vm07 bash[17031]: audit 2026-03-07T06:59:50.372813+0000 mon.vm07 (mon.0) 846 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:50.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:50 vm07 bash[17031]: audit 2026-03-07T06:59:50.372813+0000 mon.vm07 (mon.0) 846 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:50.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:50 vm09 bash[21220]: cluster 2026-03-07T06:59:48.724456+0000 mgr.vm07.yrfcuj (mgr.14201) 290 : cluster [DBG] pgmap v182: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:50 vm09 bash[21220]: cluster 2026-03-07T06:59:48.724456+0000 mgr.vm07.yrfcuj (mgr.14201) 290 : cluster [DBG] pgmap v182: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T07:59:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:50 vm09 bash[21220]: audit 2026-03-07T06:59:50.012927+0000 mon.vm07 (mon.0) 843 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:50 vm09 bash[21220]: audit 2026-03-07T06:59:50.012927+0000 mon.vm07 (mon.0) 843 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:50 vm09 bash[21220]: audit 2026-03-07T06:59:50.019498+0000 mon.vm07 (mon.0) 844 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:50 vm09 bash[21220]: audit 2026-03-07T06:59:50.019498+0000 mon.vm07 (mon.0) 844 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:50 vm09 bash[21220]: audit 2026-03-07T06:59:50.366273+0000 mon.vm07 (mon.0) 845 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:50 vm09 bash[21220]: audit 2026-03-07T06:59:50.366273+0000 mon.vm07 (mon.0) 845 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:50 vm09 bash[21220]: audit 2026-03-07T06:59:50.372813+0000 mon.vm07 (mon.0) 846 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:50 vm09 bash[21220]: audit 2026-03-07T06:59:50.372813+0000 mon.vm07 (mon.0) 846 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:51.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:51 vm09 bash[21220]: audit 2026-03-07T06:59:50.681674+0000 mon.vm07 (mon.0) 847 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:59:51.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:51 vm09 bash[21220]: audit 2026-03-07T06:59:50.681674+0000 mon.vm07 (mon.0) 847 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:59:51.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:51 vm09 bash[21220]: audit 2026-03-07T06:59:50.682257+0000 mon.vm07 (mon.0) 848 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:59:51.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:51 vm09 bash[21220]: audit 2026-03-07T06:59:50.682257+0000 mon.vm07 (mon.0) 848 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:59:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:51 vm09 bash[21220]: audit 2026-03-07T06:59:50.687425+0000 mon.vm07 (mon.0) 849 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:51 vm09 bash[21220]: audit 2026-03-07T06:59:50.687425+0000 mon.vm07 (mon.0) 849 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:51 vm09 bash[21220]: audit 2026-03-07T06:59:50.688828+0000 mon.vm07 (mon.0) 850 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:59:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:51 vm09 bash[21220]: audit 2026-03-07T06:59:50.688828+0000 mon.vm07 (mon.0) 850 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:59:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:51 vm07 bash[17031]: audit 2026-03-07T06:59:50.681674+0000 mon.vm07 (mon.0) 847 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:59:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:51 vm07 bash[17031]: audit 2026-03-07T06:59:50.681674+0000 mon.vm07 (mon.0) 847 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T07:59:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:51 vm07 bash[17031]: audit 2026-03-07T06:59:50.682257+0000 mon.vm07 (mon.0) 848 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:59:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:51 vm07 bash[17031]: audit 2026-03-07T06:59:50.682257+0000 mon.vm07 (mon.0) 848 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T07:59:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:51 vm07 bash[17031]: audit 2026-03-07T06:59:50.687425+0000 mon.vm07 (mon.0) 849 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:51 vm07 bash[17031]: audit 2026-03-07T06:59:50.687425+0000 mon.vm07 (mon.0) 849 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T07:59:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:51 vm07 bash[17031]: audit 2026-03-07T06:59:50.688828+0000 mon.vm07 (mon.0) 850 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:59:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:51 vm07 bash[17031]: audit 2026-03-07T06:59:50.688828+0000 mon.vm07 (mon.0) 850 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T07:59:52.824 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T07:59:52.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:52 vm09 bash[21220]: cluster 2026-03-07T06:59:50.683118+0000 mgr.vm07.yrfcuj (mgr.14201) 291 : cluster [DBG] pgmap v183: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 256 B/s rd, 513 B/s wr, 0 op/s 2026-03-07T07:59:52.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:52 vm09 bash[21220]: cluster 2026-03-07T06:59:50.683118+0000 mgr.vm07.yrfcuj (mgr.14201) 291 : cluster [DBG] pgmap v183: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 256 B/s rd, 513 B/s wr, 0 op/s 2026-03-07T07:59:52.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:52 vm09 bash[21220]: cluster 2026-03-07T06:59:50.683280+0000 mgr.vm07.yrfcuj (mgr.14201) 292 : cluster [DBG] pgmap v184: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 308 B/s rd, 616 B/s wr, 0 op/s 2026-03-07T07:59:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:52 vm09 bash[21220]: cluster 2026-03-07T06:59:50.683280+0000 mgr.vm07.yrfcuj (mgr.14201) 292 : cluster [DBG] pgmap v184: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 308 B/s rd, 616 B/s wr, 0 op/s 2026-03-07T07:59:52.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:52 vm07 bash[17031]: cluster 2026-03-07T06:59:50.683118+0000 mgr.vm07.yrfcuj (mgr.14201) 291 : cluster [DBG] pgmap v183: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 256 B/s rd, 513 B/s wr, 0 op/s 2026-03-07T07:59:52.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:52 vm07 bash[17031]: cluster 2026-03-07T06:59:50.683118+0000 mgr.vm07.yrfcuj (mgr.14201) 291 : cluster [DBG] pgmap v183: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 256 B/s rd, 513 B/s wr, 0 op/s 2026-03-07T07:59:52.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:52 vm07 bash[17031]: cluster 2026-03-07T06:59:50.683280+0000 mgr.vm07.yrfcuj (mgr.14201) 292 : cluster [DBG] pgmap v184: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 308 B/s rd, 616 B/s wr, 0 op/s 2026-03-07T07:59:52.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:52 vm07 bash[17031]: cluster 2026-03-07T06:59:50.683280+0000 mgr.vm07.yrfcuj (mgr.14201) 292 : cluster [DBG] pgmap v184: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 308 B/s rd, 616 B/s wr, 0 op/s 2026-03-07T07:59:52.979 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:59:52.979 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 2s ago 2m - - 2026-03-07T07:59:52.979 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 2s ago 2m 100M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T07:59:52.979 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 2s ago 2m 95.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T07:59:52.979 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (2m) 2s ago 2m 95.8M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T07:59:53.170 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T07:59:53.171 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T07:59:53.171 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T07:59:53.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:53 vm07 bash[17031]: audit 2026-03-07T06:59:53.170017+0000 mon.vm09 (mon.1) 31 : audit [DBG] from='client.? 192.168.123.107:0/2732921288' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:53.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:53 vm07 bash[17031]: audit 2026-03-07T06:59:53.170017+0000 mon.vm09 (mon.1) 31 : audit [DBG] from='client.? 192.168.123.107:0/2732921288' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:53.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:53 vm09 bash[21220]: audit 2026-03-07T06:59:53.170017+0000 mon.vm09 (mon.1) 31 : audit [DBG] from='client.? 192.168.123.107:0/2732921288' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:53.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:53 vm09 bash[21220]: audit 2026-03-07T06:59:53.170017+0000 mon.vm09 (mon.1) 31 : audit [DBG] from='client.? 192.168.123.107:0/2732921288' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:54.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:54 vm09 bash[21220]: cluster 2026-03-07T06:59:52.683721+0000 mgr.vm07.yrfcuj (mgr.14201) 293 : cluster [DBG] pgmap v185: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 308 B/s rd, 616 B/s wr, 0 op/s 2026-03-07T07:59:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:54 vm09 bash[21220]: cluster 2026-03-07T06:59:52.683721+0000 mgr.vm07.yrfcuj (mgr.14201) 293 : cluster [DBG] pgmap v185: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 308 B/s rd, 616 B/s wr, 0 op/s 2026-03-07T07:59:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:54 vm09 bash[21220]: audit 2026-03-07T06:59:52.812996+0000 mgr.vm07.yrfcuj (mgr.14201) 294 : audit [DBG] from='client.24527 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:54 vm09 bash[21220]: audit 2026-03-07T06:59:52.812996+0000 mgr.vm07.yrfcuj (mgr.14201) 294 : audit [DBG] from='client.24527 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:54 vm09 bash[21220]: audit 2026-03-07T06:59:52.981108+0000 mgr.vm07.yrfcuj (mgr.14201) 295 : audit [DBG] from='client.14836 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:54 vm09 bash[21220]: audit 2026-03-07T06:59:52.981108+0000 mgr.vm07.yrfcuj (mgr.14201) 295 : audit [DBG] from='client.14836 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:54 vm07 bash[17031]: cluster 2026-03-07T06:59:52.683721+0000 mgr.vm07.yrfcuj (mgr.14201) 293 : cluster [DBG] pgmap v185: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 308 B/s rd, 616 B/s wr, 0 op/s 2026-03-07T07:59:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:54 vm07 bash[17031]: cluster 2026-03-07T06:59:52.683721+0000 mgr.vm07.yrfcuj (mgr.14201) 293 : cluster [DBG] pgmap v185: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 308 B/s rd, 616 B/s wr, 0 op/s 2026-03-07T07:59:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:54 vm07 bash[17031]: audit 2026-03-07T06:59:52.812996+0000 mgr.vm07.yrfcuj (mgr.14201) 294 : audit [DBG] from='client.24527 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:54 vm07 bash[17031]: audit 2026-03-07T06:59:52.812996+0000 mgr.vm07.yrfcuj (mgr.14201) 294 : audit [DBG] from='client.24527 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:54 vm07 bash[17031]: audit 2026-03-07T06:59:52.981108+0000 mgr.vm07.yrfcuj (mgr.14201) 295 : audit [DBG] from='client.14836 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:54 vm07 bash[17031]: audit 2026-03-07T06:59:52.981108+0000 mgr.vm07.yrfcuj (mgr.14201) 295 : audit [DBG] from='client.14836 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:55.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:55 vm09 bash[21220]: audit 2026-03-07T06:59:55.129165+0000 mon.vm07 (mon.0) 851 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:59:55.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:55 vm09 bash[21220]: audit 2026-03-07T06:59:55.129165+0000 mon.vm07 (mon.0) 851 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:59:55.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:55 vm07 bash[17031]: audit 2026-03-07T06:59:55.129165+0000 mon.vm07 (mon.0) 851 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:59:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:55 vm07 bash[17031]: audit 2026-03-07T06:59:55.129165+0000 mon.vm07 (mon.0) 851 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T07:59:56.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:56 vm09 bash[21220]: cluster 2026-03-07T06:59:54.684077+0000 mgr.vm07.yrfcuj (mgr.14201) 296 : cluster [DBG] pgmap v186: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 308 B/s rd, 616 B/s wr, 0 op/s 2026-03-07T07:59:56.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:56 vm09 bash[21220]: cluster 2026-03-07T06:59:54.684077+0000 mgr.vm07.yrfcuj (mgr.14201) 296 : cluster [DBG] pgmap v186: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 308 B/s rd, 616 B/s wr, 0 op/s 2026-03-07T07:59:56.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:56 vm07 bash[17031]: cluster 2026-03-07T06:59:54.684077+0000 mgr.vm07.yrfcuj (mgr.14201) 296 : cluster [DBG] pgmap v186: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 308 B/s rd, 616 B/s wr, 0 op/s 2026-03-07T07:59:56.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:56 vm07 bash[17031]: cluster 2026-03-07T06:59:54.684077+0000 mgr.vm07.yrfcuj (mgr.14201) 296 : cluster [DBG] pgmap v186: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 308 B/s rd, 616 B/s wr, 0 op/s 2026-03-07T07:59:58.362 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T07:59:58.526 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T07:59:58.526 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 8s ago 2m - - 2026-03-07T07:59:58.526 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 8s ago 2m 100M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T07:59:58.526 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 8s ago 2m 95.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T07:59:58.526 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (2m) 8s ago 2m 95.8M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T07:59:58.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:58 vm07 bash[17031]: cluster 2026-03-07T06:59:56.684427+0000 mgr.vm07.yrfcuj (mgr.14201) 297 : cluster [DBG] pgmap v187: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:58.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:58 vm07 bash[17031]: cluster 2026-03-07T06:59:56.684427+0000 mgr.vm07.yrfcuj (mgr.14201) 297 : cluster [DBG] pgmap v187: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:58.720 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T07:59:58.720 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T07:59:58.720 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T07:59:58.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:58 vm09 bash[21220]: cluster 2026-03-07T06:59:56.684427+0000 mgr.vm07.yrfcuj (mgr.14201) 297 : cluster [DBG] pgmap v187: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:58.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:58 vm09 bash[21220]: cluster 2026-03-07T06:59:56.684427+0000 mgr.vm07.yrfcuj (mgr.14201) 297 : cluster [DBG] pgmap v187: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T07:59:59.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:59 vm09 bash[21220]: audit 2026-03-07T06:59:58.345742+0000 mgr.vm07.yrfcuj (mgr.14201) 298 : audit [DBG] from='client.24535 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:59.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:59 vm09 bash[21220]: audit 2026-03-07T06:59:58.345742+0000 mgr.vm07.yrfcuj (mgr.14201) 298 : audit [DBG] from='client.24535 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:59.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:59 vm09 bash[21220]: audit 2026-03-07T06:59:58.725214+0000 mon.vm07 (mon.0) 852 : audit [DBG] from='client.? 192.168.123.107:0/2458022880' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:59.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 07:59:59 vm09 bash[21220]: audit 2026-03-07T06:59:58.725214+0000 mon.vm07 (mon.0) 852 : audit [DBG] from='client.? 192.168.123.107:0/2458022880' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:59.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:59 vm07 bash[17031]: audit 2026-03-07T06:59:58.345742+0000 mgr.vm07.yrfcuj (mgr.14201) 298 : audit [DBG] from='client.24535 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:59.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:59 vm07 bash[17031]: audit 2026-03-07T06:59:58.345742+0000 mgr.vm07.yrfcuj (mgr.14201) 298 : audit [DBG] from='client.24535 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T07:59:59.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:59 vm07 bash[17031]: audit 2026-03-07T06:59:58.725214+0000 mon.vm07 (mon.0) 852 : audit [DBG] from='client.? 192.168.123.107:0/2458022880' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T07:59:59.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 07:59:59 vm07 bash[17031]: audit 2026-03-07T06:59:58.725214+0000 mon.vm07 (mon.0) 852 : audit [DBG] from='client.? 192.168.123.107:0/2458022880' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:00.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:00 vm09 bash[21220]: audit 2026-03-07T06:59:58.528644+0000 mgr.vm07.yrfcuj (mgr.14201) 299 : audit [DBG] from='client.14848 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:00.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:00 vm09 bash[21220]: audit 2026-03-07T06:59:58.528644+0000 mgr.vm07.yrfcuj (mgr.14201) 299 : audit [DBG] from='client.14848 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:00 vm09 bash[21220]: cluster 2026-03-07T06:59:58.684917+0000 mgr.vm07.yrfcuj (mgr.14201) 300 : cluster [DBG] pgmap v188: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:00 vm09 bash[21220]: cluster 2026-03-07T06:59:58.684917+0000 mgr.vm07.yrfcuj (mgr.14201) 300 : cluster [DBG] pgmap v188: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:00 vm09 bash[21220]: cluster 2026-03-07T07:00:00.000160+0000 mon.vm07 (mon.0) 853 : cluster [WRN] Health detail: HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:00:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:00 vm09 bash[21220]: cluster 2026-03-07T07:00:00.000160+0000 mon.vm07 (mon.0) 853 : cluster [WRN] Health detail: HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:00:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:00 vm09 bash[21220]: cluster 2026-03-07T07:00:00.000184+0000 mon.vm07 (mon.0) 854 : cluster [WRN] [WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:00:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:00 vm09 bash[21220]: cluster 2026-03-07T07:00:00.000184+0000 mon.vm07 (mon.0) 854 : cluster [WRN] [WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:00:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:00 vm09 bash[21220]: cluster 2026-03-07T07:00:00.000189+0000 mon.vm07 (mon.0) 855 : cluster [WRN] daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:00:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:00 vm09 bash[21220]: cluster 2026-03-07T07:00:00.000189+0000 mon.vm07 (mon.0) 855 : cluster [WRN] daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:00:00.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:00 vm07 bash[17031]: audit 2026-03-07T06:59:58.528644+0000 mgr.vm07.yrfcuj (mgr.14201) 299 : audit [DBG] from='client.14848 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:00.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:00 vm07 bash[17031]: audit 2026-03-07T06:59:58.528644+0000 mgr.vm07.yrfcuj (mgr.14201) 299 : audit [DBG] from='client.14848 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:00.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:00 vm07 bash[17031]: cluster 2026-03-07T06:59:58.684917+0000 mgr.vm07.yrfcuj (mgr.14201) 300 : cluster [DBG] pgmap v188: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:00.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:00 vm07 bash[17031]: cluster 2026-03-07T06:59:58.684917+0000 mgr.vm07.yrfcuj (mgr.14201) 300 : cluster [DBG] pgmap v188: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:00.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:00 vm07 bash[17031]: cluster 2026-03-07T07:00:00.000160+0000 mon.vm07 (mon.0) 853 : cluster [WRN] Health detail: HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:00:00.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:00 vm07 bash[17031]: cluster 2026-03-07T07:00:00.000160+0000 mon.vm07 (mon.0) 853 : cluster [WRN] Health detail: HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:00:00.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:00 vm07 bash[17031]: cluster 2026-03-07T07:00:00.000184+0000 mon.vm07 (mon.0) 854 : cluster [WRN] [WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:00:00.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:00 vm07 bash[17031]: cluster 2026-03-07T07:00:00.000184+0000 mon.vm07 (mon.0) 854 : cluster [WRN] [WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:00:00.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:00 vm07 bash[17031]: cluster 2026-03-07T07:00:00.000189+0000 mon.vm07 (mon.0) 855 : cluster [WRN] daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:00:00.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:00 vm07 bash[17031]: cluster 2026-03-07T07:00:00.000189+0000 mon.vm07 (mon.0) 855 : cluster [WRN] daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:00:02.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:02 vm09 bash[21220]: cluster 2026-03-07T07:00:00.685461+0000 mgr.vm07.yrfcuj (mgr.14201) 301 : cluster [DBG] pgmap v189: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:02.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:02 vm09 bash[21220]: cluster 2026-03-07T07:00:00.685461+0000 mgr.vm07.yrfcuj (mgr.14201) 301 : cluster [DBG] pgmap v189: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:02.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:02 vm07 bash[17031]: cluster 2026-03-07T07:00:00.685461+0000 mgr.vm07.yrfcuj (mgr.14201) 301 : cluster [DBG] pgmap v189: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:02.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:02 vm07 bash[17031]: cluster 2026-03-07T07:00:00.685461+0000 mgr.vm07.yrfcuj (mgr.14201) 301 : cluster [DBG] pgmap v189: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:03.907 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:00:04.069 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:00:04.070 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 13s ago 2m - - 2026-03-07T08:00:04.070 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 13s ago 2m 100M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:00:04.070 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 14s ago 2m 95.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:00:04.070 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (2m) 14s ago 2m 95.8M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:00:04.254 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:00:04.254 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:00:04.254 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:00:04.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:04 vm09 bash[21220]: cluster 2026-03-07T07:00:02.685992+0000 mgr.vm07.yrfcuj (mgr.14201) 302 : cluster [DBG] pgmap v190: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:04.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:04 vm09 bash[21220]: cluster 2026-03-07T07:00:02.685992+0000 mgr.vm07.yrfcuj (mgr.14201) 302 : cluster [DBG] pgmap v190: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:04.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:04 vm09 bash[21220]: audit 2026-03-07T07:00:04.259741+0000 mon.vm07 (mon.0) 856 : audit [DBG] from='client.? 192.168.123.107:0/2902415490' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:04.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:04 vm09 bash[21220]: audit 2026-03-07T07:00:04.259741+0000 mon.vm07 (mon.0) 856 : audit [DBG] from='client.? 192.168.123.107:0/2902415490' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:04.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:04 vm07 bash[17031]: cluster 2026-03-07T07:00:02.685992+0000 mgr.vm07.yrfcuj (mgr.14201) 302 : cluster [DBG] pgmap v190: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:04.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:04 vm07 bash[17031]: cluster 2026-03-07T07:00:02.685992+0000 mgr.vm07.yrfcuj (mgr.14201) 302 : cluster [DBG] pgmap v190: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:04.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:04 vm07 bash[17031]: audit 2026-03-07T07:00:04.259741+0000 mon.vm07 (mon.0) 856 : audit [DBG] from='client.? 192.168.123.107:0/2902415490' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:04.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:04 vm07 bash[17031]: audit 2026-03-07T07:00:04.259741+0000 mon.vm07 (mon.0) 856 : audit [DBG] from='client.? 192.168.123.107:0/2902415490' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:05 vm09 bash[21220]: audit 2026-03-07T07:00:03.896454+0000 mgr.vm07.yrfcuj (mgr.14201) 303 : audit [DBG] from='client.14856 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:05 vm09 bash[21220]: audit 2026-03-07T07:00:03.896454+0000 mgr.vm07.yrfcuj (mgr.14201) 303 : audit [DBG] from='client.14856 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:05 vm09 bash[21220]: audit 2026-03-07T07:00:04.072327+0000 mgr.vm07.yrfcuj (mgr.14201) 304 : audit [DBG] from='client.14860 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:05 vm09 bash[21220]: audit 2026-03-07T07:00:04.072327+0000 mgr.vm07.yrfcuj (mgr.14201) 304 : audit [DBG] from='client.14860 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:05.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:05 vm07 bash[17031]: audit 2026-03-07T07:00:03.896454+0000 mgr.vm07.yrfcuj (mgr.14201) 303 : audit [DBG] from='client.14856 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:05.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:05 vm07 bash[17031]: audit 2026-03-07T07:00:03.896454+0000 mgr.vm07.yrfcuj (mgr.14201) 303 : audit [DBG] from='client.14856 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:05.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:05 vm07 bash[17031]: audit 2026-03-07T07:00:04.072327+0000 mgr.vm07.yrfcuj (mgr.14201) 304 : audit [DBG] from='client.14860 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:05.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:05 vm07 bash[17031]: audit 2026-03-07T07:00:04.072327+0000 mgr.vm07.yrfcuj (mgr.14201) 304 : audit [DBG] from='client.14860 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:06 vm09 bash[21220]: cluster 2026-03-07T07:00:04.686434+0000 mgr.vm07.yrfcuj (mgr.14201) 305 : cluster [DBG] pgmap v191: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:06 vm09 bash[21220]: cluster 2026-03-07T07:00:04.686434+0000 mgr.vm07.yrfcuj (mgr.14201) 305 : cluster [DBG] pgmap v191: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:06.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:06 vm07 bash[17031]: cluster 2026-03-07T07:00:04.686434+0000 mgr.vm07.yrfcuj (mgr.14201) 305 : cluster [DBG] pgmap v191: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:06.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:06 vm07 bash[17031]: cluster 2026-03-07T07:00:04.686434+0000 mgr.vm07.yrfcuj (mgr.14201) 305 : cluster [DBG] pgmap v191: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:08.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:08 vm09 bash[21220]: cluster 2026-03-07T07:00:06.686842+0000 mgr.vm07.yrfcuj (mgr.14201) 306 : cluster [DBG] pgmap v192: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:08.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:08 vm09 bash[21220]: cluster 2026-03-07T07:00:06.686842+0000 mgr.vm07.yrfcuj (mgr.14201) 306 : cluster [DBG] pgmap v192: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:08.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:08 vm07 bash[17031]: cluster 2026-03-07T07:00:06.686842+0000 mgr.vm07.yrfcuj (mgr.14201) 306 : cluster [DBG] pgmap v192: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:08.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:08 vm07 bash[17031]: cluster 2026-03-07T07:00:06.686842+0000 mgr.vm07.yrfcuj (mgr.14201) 306 : cluster [DBG] pgmap v192: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:09.436 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:00:09.600 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:00:09.600 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 19s ago 2m - - 2026-03-07T08:00:09.600 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 19s ago 2m 100M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:00:09.600 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 19s ago 2m 95.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:00:09.600 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (2m) 19s ago 2m 95.8M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:00:09.790 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:00:09.791 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:00:09.791 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:00:09.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:09 vm09 bash[21220]: cluster 2026-03-07T07:00:08.687266+0000 mgr.vm07.yrfcuj (mgr.14201) 307 : cluster [DBG] pgmap v193: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:09.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:09 vm09 bash[21220]: cluster 2026-03-07T07:00:08.687266+0000 mgr.vm07.yrfcuj (mgr.14201) 307 : cluster [DBG] pgmap v193: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:09.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:09 vm09 bash[21220]: audit 2026-03-07T07:00:09.425210+0000 mgr.vm07.yrfcuj (mgr.14201) 308 : audit [DBG] from='client.14868 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:09.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:09 vm09 bash[21220]: audit 2026-03-07T07:00:09.425210+0000 mgr.vm07.yrfcuj (mgr.14201) 308 : audit [DBG] from='client.14868 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:09.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:09 vm07 bash[17031]: cluster 2026-03-07T07:00:08.687266+0000 mgr.vm07.yrfcuj (mgr.14201) 307 : cluster [DBG] pgmap v193: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:09.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:09 vm07 bash[17031]: cluster 2026-03-07T07:00:08.687266+0000 mgr.vm07.yrfcuj (mgr.14201) 307 : cluster [DBG] pgmap v193: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:09.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:09 vm07 bash[17031]: audit 2026-03-07T07:00:09.425210+0000 mgr.vm07.yrfcuj (mgr.14201) 308 : audit [DBG] from='client.14868 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:09.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:09 vm07 bash[17031]: audit 2026-03-07T07:00:09.425210+0000 mgr.vm07.yrfcuj (mgr.14201) 308 : audit [DBG] from='client.14868 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:10 vm09 bash[21220]: audit 2026-03-07T07:00:09.602822+0000 mgr.vm07.yrfcuj (mgr.14201) 309 : audit [DBG] from='client.14872 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:10 vm09 bash[21220]: audit 2026-03-07T07:00:09.602822+0000 mgr.vm07.yrfcuj (mgr.14201) 309 : audit [DBG] from='client.14872 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:10 vm09 bash[21220]: audit 2026-03-07T07:00:09.789621+0000 mon.vm09 (mon.1) 32 : audit [DBG] from='client.? 192.168.123.107:0/1891130255' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:10 vm09 bash[21220]: audit 2026-03-07T07:00:09.789621+0000 mon.vm09 (mon.1) 32 : audit [DBG] from='client.? 192.168.123.107:0/1891130255' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:10 vm09 bash[21220]: audit 2026-03-07T07:00:10.129497+0000 mon.vm07 (mon.0) 857 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:00:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:10 vm09 bash[21220]: audit 2026-03-07T07:00:10.129497+0000 mon.vm07 (mon.0) 857 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:00:10.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:10 vm07 bash[17031]: audit 2026-03-07T07:00:09.602822+0000 mgr.vm07.yrfcuj (mgr.14201) 309 : audit [DBG] from='client.14872 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:10.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:10 vm07 bash[17031]: audit 2026-03-07T07:00:09.602822+0000 mgr.vm07.yrfcuj (mgr.14201) 309 : audit [DBG] from='client.14872 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:10.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:10 vm07 bash[17031]: audit 2026-03-07T07:00:09.789621+0000 mon.vm09 (mon.1) 32 : audit [DBG] from='client.? 192.168.123.107:0/1891130255' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:10.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:10 vm07 bash[17031]: audit 2026-03-07T07:00:09.789621+0000 mon.vm09 (mon.1) 32 : audit [DBG] from='client.? 192.168.123.107:0/1891130255' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:10.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:10 vm07 bash[17031]: audit 2026-03-07T07:00:10.129497+0000 mon.vm07 (mon.0) 857 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:00:10.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:10 vm07 bash[17031]: audit 2026-03-07T07:00:10.129497+0000 mon.vm07 (mon.0) 857 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:00:11.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:11 vm09 bash[21220]: cluster 2026-03-07T07:00:10.687707+0000 mgr.vm07.yrfcuj (mgr.14201) 310 : cluster [DBG] pgmap v194: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:11.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:11 vm09 bash[21220]: cluster 2026-03-07T07:00:10.687707+0000 mgr.vm07.yrfcuj (mgr.14201) 310 : cluster [DBG] pgmap v194: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:11.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:11 vm07 bash[17031]: cluster 2026-03-07T07:00:10.687707+0000 mgr.vm07.yrfcuj (mgr.14201) 310 : cluster [DBG] pgmap v194: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:11.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:11 vm07 bash[17031]: cluster 2026-03-07T07:00:10.687707+0000 mgr.vm07.yrfcuj (mgr.14201) 310 : cluster [DBG] pgmap v194: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:14.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:13 vm09 bash[21220]: cluster 2026-03-07T07:00:12.688208+0000 mgr.vm07.yrfcuj (mgr.14201) 311 : cluster [DBG] pgmap v195: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:14.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:13 vm09 bash[21220]: cluster 2026-03-07T07:00:12.688208+0000 mgr.vm07.yrfcuj (mgr.14201) 311 : cluster [DBG] pgmap v195: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:14.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:13 vm07 bash[17031]: cluster 2026-03-07T07:00:12.688208+0000 mgr.vm07.yrfcuj (mgr.14201) 311 : cluster [DBG] pgmap v195: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:14.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:13 vm07 bash[17031]: cluster 2026-03-07T07:00:12.688208+0000 mgr.vm07.yrfcuj (mgr.14201) 311 : cluster [DBG] pgmap v195: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:14.967 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:00:15.115 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:00:15.115 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 24s ago 2m - - 2026-03-07T08:00:15.115 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 24s ago 2m 100M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:00:15.115 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 25s ago 2m 95.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:00:15.115 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (2m) 25s ago 2m 95.8M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:00:15.323 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:00:15.323 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:00:15.323 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:00:16.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:15 vm09 bash[21220]: cluster 2026-03-07T07:00:14.688602+0000 mgr.vm07.yrfcuj (mgr.14201) 312 : cluster [DBG] pgmap v196: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:16.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:15 vm09 bash[21220]: cluster 2026-03-07T07:00:14.688602+0000 mgr.vm07.yrfcuj (mgr.14201) 312 : cluster [DBG] pgmap v196: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:16.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:15 vm09 bash[21220]: audit 2026-03-07T07:00:14.956315+0000 mgr.vm07.yrfcuj (mgr.14201) 313 : audit [DBG] from='client.14880 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:16.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:15 vm09 bash[21220]: audit 2026-03-07T07:00:14.956315+0000 mgr.vm07.yrfcuj (mgr.14201) 313 : audit [DBG] from='client.14880 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:16.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:15 vm09 bash[21220]: audit 2026-03-07T07:00:15.118670+0000 mgr.vm07.yrfcuj (mgr.14201) 314 : audit [DBG] from='client.14884 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:16.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:15 vm09 bash[21220]: audit 2026-03-07T07:00:15.118670+0000 mgr.vm07.yrfcuj (mgr.14201) 314 : audit [DBG] from='client.14884 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:16.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:15 vm09 bash[21220]: audit 2026-03-07T07:00:15.328415+0000 mon.vm07 (mon.0) 858 : audit [DBG] from='client.? 192.168.123.107:0/861703237' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:16.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:15 vm09 bash[21220]: audit 2026-03-07T07:00:15.328415+0000 mon.vm07 (mon.0) 858 : audit [DBG] from='client.? 192.168.123.107:0/861703237' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:16.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:15 vm07 bash[17031]: cluster 2026-03-07T07:00:14.688602+0000 mgr.vm07.yrfcuj (mgr.14201) 312 : cluster [DBG] pgmap v196: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:16.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:15 vm07 bash[17031]: cluster 2026-03-07T07:00:14.688602+0000 mgr.vm07.yrfcuj (mgr.14201) 312 : cluster [DBG] pgmap v196: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:16.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:15 vm07 bash[17031]: audit 2026-03-07T07:00:14.956315+0000 mgr.vm07.yrfcuj (mgr.14201) 313 : audit [DBG] from='client.14880 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:16.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:15 vm07 bash[17031]: audit 2026-03-07T07:00:14.956315+0000 mgr.vm07.yrfcuj (mgr.14201) 313 : audit [DBG] from='client.14880 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:16.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:15 vm07 bash[17031]: audit 2026-03-07T07:00:15.118670+0000 mgr.vm07.yrfcuj (mgr.14201) 314 : audit [DBG] from='client.14884 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:16.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:15 vm07 bash[17031]: audit 2026-03-07T07:00:15.118670+0000 mgr.vm07.yrfcuj (mgr.14201) 314 : audit [DBG] from='client.14884 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:16.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:15 vm07 bash[17031]: audit 2026-03-07T07:00:15.328415+0000 mon.vm07 (mon.0) 858 : audit [DBG] from='client.? 192.168.123.107:0/861703237' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:16.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:15 vm07 bash[17031]: audit 2026-03-07T07:00:15.328415+0000 mon.vm07 (mon.0) 858 : audit [DBG] from='client.? 192.168.123.107:0/861703237' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:18.079 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:17 vm09 bash[21220]: cluster 2026-03-07T07:00:16.688915+0000 mgr.vm07.yrfcuj (mgr.14201) 315 : cluster [DBG] pgmap v197: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:18.079 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:17 vm09 bash[21220]: cluster 2026-03-07T07:00:16.688915+0000 mgr.vm07.yrfcuj (mgr.14201) 315 : cluster [DBG] pgmap v197: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:18.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:17 vm07 bash[17031]: cluster 2026-03-07T07:00:16.688915+0000 mgr.vm07.yrfcuj (mgr.14201) 315 : cluster [DBG] pgmap v197: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:18.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:17 vm07 bash[17031]: cluster 2026-03-07T07:00:16.688915+0000 mgr.vm07.yrfcuj (mgr.14201) 315 : cluster [DBG] pgmap v197: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:20.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:19 vm09 bash[21220]: cluster 2026-03-07T07:00:18.689364+0000 mgr.vm07.yrfcuj (mgr.14201) 316 : cluster [DBG] pgmap v198: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:20.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:19 vm09 bash[21220]: cluster 2026-03-07T07:00:18.689364+0000 mgr.vm07.yrfcuj (mgr.14201) 316 : cluster [DBG] pgmap v198: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:20.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:19 vm07 bash[17031]: cluster 2026-03-07T07:00:18.689364+0000 mgr.vm07.yrfcuj (mgr.14201) 316 : cluster [DBG] pgmap v198: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:20.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:19 vm07 bash[17031]: cluster 2026-03-07T07:00:18.689364+0000 mgr.vm07.yrfcuj (mgr.14201) 316 : cluster [DBG] pgmap v198: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:20.505 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:00:20.661 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:00:20.661 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 30s ago 2m - - 2026-03-07T08:00:20.661 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 30s ago 2m 100M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:00:20.661 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 30s ago 2m 95.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:00:20.661 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (2m) 30s ago 2m 95.8M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:00:20.857 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:00:20.857 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:00:20.858 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:00:21.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:20 vm09 bash[21220]: audit 2026-03-07T07:00:20.492667+0000 mgr.vm07.yrfcuj (mgr.14201) 317 : audit [DBG] from='client.14892 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:21.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:20 vm09 bash[21220]: audit 2026-03-07T07:00:20.492667+0000 mgr.vm07.yrfcuj (mgr.14201) 317 : audit [DBG] from='client.14892 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:21.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:20 vm07 bash[17031]: audit 2026-03-07T07:00:20.492667+0000 mgr.vm07.yrfcuj (mgr.14201) 317 : audit [DBG] from='client.14892 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:21.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:20 vm07 bash[17031]: audit 2026-03-07T07:00:20.492667+0000 mgr.vm07.yrfcuj (mgr.14201) 317 : audit [DBG] from='client.14892 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:22.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:21 vm09 bash[21220]: audit 2026-03-07T07:00:20.663978+0000 mgr.vm07.yrfcuj (mgr.14201) 318 : audit [DBG] from='client.14896 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:22.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:21 vm09 bash[21220]: audit 2026-03-07T07:00:20.663978+0000 mgr.vm07.yrfcuj (mgr.14201) 318 : audit [DBG] from='client.14896 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:22.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:21 vm09 bash[21220]: cluster 2026-03-07T07:00:20.689736+0000 mgr.vm07.yrfcuj (mgr.14201) 319 : cluster [DBG] pgmap v199: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:22.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:21 vm09 bash[21220]: cluster 2026-03-07T07:00:20.689736+0000 mgr.vm07.yrfcuj (mgr.14201) 319 : cluster [DBG] pgmap v199: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:22.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:21 vm09 bash[21220]: audit 2026-03-07T07:00:20.862595+0000 mon.vm07 (mon.0) 859 : audit [DBG] from='client.? 192.168.123.107:0/1444848411' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:22.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:21 vm09 bash[21220]: audit 2026-03-07T07:00:20.862595+0000 mon.vm07 (mon.0) 859 : audit [DBG] from='client.? 192.168.123.107:0/1444848411' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:22.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:21 vm07 bash[17031]: audit 2026-03-07T07:00:20.663978+0000 mgr.vm07.yrfcuj (mgr.14201) 318 : audit [DBG] from='client.14896 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:22.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:21 vm07 bash[17031]: audit 2026-03-07T07:00:20.663978+0000 mgr.vm07.yrfcuj (mgr.14201) 318 : audit [DBG] from='client.14896 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:22.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:21 vm07 bash[17031]: cluster 2026-03-07T07:00:20.689736+0000 mgr.vm07.yrfcuj (mgr.14201) 319 : cluster [DBG] pgmap v199: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:22.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:21 vm07 bash[17031]: cluster 2026-03-07T07:00:20.689736+0000 mgr.vm07.yrfcuj (mgr.14201) 319 : cluster [DBG] pgmap v199: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:22.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:21 vm07 bash[17031]: audit 2026-03-07T07:00:20.862595+0000 mon.vm07 (mon.0) 859 : audit [DBG] from='client.? 192.168.123.107:0/1444848411' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:22.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:21 vm07 bash[17031]: audit 2026-03-07T07:00:20.862595+0000 mon.vm07 (mon.0) 859 : audit [DBG] from='client.? 192.168.123.107:0/1444848411' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:23.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:23 vm07 bash[17031]: cluster 2026-03-07T07:00:22.690245+0000 mgr.vm07.yrfcuj (mgr.14201) 320 : cluster [DBG] pgmap v200: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:23.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:23 vm07 bash[17031]: cluster 2026-03-07T07:00:22.690245+0000 mgr.vm07.yrfcuj (mgr.14201) 320 : cluster [DBG] pgmap v200: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:24.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:23 vm09 bash[21220]: cluster 2026-03-07T07:00:22.690245+0000 mgr.vm07.yrfcuj (mgr.14201) 320 : cluster [DBG] pgmap v200: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:24.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:23 vm09 bash[21220]: cluster 2026-03-07T07:00:22.690245+0000 mgr.vm07.yrfcuj (mgr.14201) 320 : cluster [DBG] pgmap v200: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:26.046 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:00:26.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:25 vm09 bash[21220]: cluster 2026-03-07T07:00:24.690594+0000 mgr.vm07.yrfcuj (mgr.14201) 321 : cluster [DBG] pgmap v201: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:26.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:25 vm09 bash[21220]: cluster 2026-03-07T07:00:24.690594+0000 mgr.vm07.yrfcuj (mgr.14201) 321 : cluster [DBG] pgmap v201: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:26.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:25 vm09 bash[21220]: audit 2026-03-07T07:00:25.129573+0000 mon.vm07 (mon.0) 860 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:00:26.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:25 vm09 bash[21220]: audit 2026-03-07T07:00:25.129573+0000 mon.vm07 (mon.0) 860 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:00:26.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:25 vm07 bash[17031]: cluster 2026-03-07T07:00:24.690594+0000 mgr.vm07.yrfcuj (mgr.14201) 321 : cluster [DBG] pgmap v201: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:26.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:25 vm07 bash[17031]: cluster 2026-03-07T07:00:24.690594+0000 mgr.vm07.yrfcuj (mgr.14201) 321 : cluster [DBG] pgmap v201: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:26.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:25 vm07 bash[17031]: audit 2026-03-07T07:00:25.129573+0000 mon.vm07 (mon.0) 860 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:00:26.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:25 vm07 bash[17031]: audit 2026-03-07T07:00:25.129573+0000 mon.vm07 (mon.0) 860 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:00:26.208 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:00:26.208 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 35s ago 2m - - 2026-03-07T08:00:26.208 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 35s ago 2m 100M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:00:26.208 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 36s ago 2m 95.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:00:26.208 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (2m) 36s ago 2m 95.8M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:00:26.403 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:00:26.403 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:00:26.403 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:00:27.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:26 vm09 bash[21220]: audit 2026-03-07T07:00:26.035042+0000 mgr.vm07.yrfcuj (mgr.14201) 322 : audit [DBG] from='client.14904 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:27.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:26 vm09 bash[21220]: audit 2026-03-07T07:00:26.035042+0000 mgr.vm07.yrfcuj (mgr.14201) 322 : audit [DBG] from='client.14904 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:27.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:26 vm09 bash[21220]: audit 2026-03-07T07:00:26.210988+0000 mgr.vm07.yrfcuj (mgr.14201) 323 : audit [DBG] from='client.14908 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:27.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:26 vm09 bash[21220]: audit 2026-03-07T07:00:26.210988+0000 mgr.vm07.yrfcuj (mgr.14201) 323 : audit [DBG] from='client.14908 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:27.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:26 vm09 bash[21220]: audit 2026-03-07T07:00:26.408034+0000 mon.vm07 (mon.0) 861 : audit [DBG] from='client.? 192.168.123.107:0/3107632289' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:27.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:26 vm09 bash[21220]: audit 2026-03-07T07:00:26.408034+0000 mon.vm07 (mon.0) 861 : audit [DBG] from='client.? 192.168.123.107:0/3107632289' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:27.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:26 vm07 bash[17031]: audit 2026-03-07T07:00:26.035042+0000 mgr.vm07.yrfcuj (mgr.14201) 322 : audit [DBG] from='client.14904 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:27.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:26 vm07 bash[17031]: audit 2026-03-07T07:00:26.035042+0000 mgr.vm07.yrfcuj (mgr.14201) 322 : audit [DBG] from='client.14904 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:27.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:26 vm07 bash[17031]: audit 2026-03-07T07:00:26.210988+0000 mgr.vm07.yrfcuj (mgr.14201) 323 : audit [DBG] from='client.14908 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:27.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:26 vm07 bash[17031]: audit 2026-03-07T07:00:26.210988+0000 mgr.vm07.yrfcuj (mgr.14201) 323 : audit [DBG] from='client.14908 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:27.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:26 vm07 bash[17031]: audit 2026-03-07T07:00:26.408034+0000 mon.vm07 (mon.0) 861 : audit [DBG] from='client.? 192.168.123.107:0/3107632289' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:27.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:26 vm07 bash[17031]: audit 2026-03-07T07:00:26.408034+0000 mon.vm07 (mon.0) 861 : audit [DBG] from='client.? 192.168.123.107:0/3107632289' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:28.079 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:27 vm09 bash[21220]: cluster 2026-03-07T07:00:26.691006+0000 mgr.vm07.yrfcuj (mgr.14201) 324 : cluster [DBG] pgmap v202: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:28.079 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:27 vm09 bash[21220]: cluster 2026-03-07T07:00:26.691006+0000 mgr.vm07.yrfcuj (mgr.14201) 324 : cluster [DBG] pgmap v202: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:28.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:27 vm07 bash[17031]: cluster 2026-03-07T07:00:26.691006+0000 mgr.vm07.yrfcuj (mgr.14201) 324 : cluster [DBG] pgmap v202: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:28.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:27 vm07 bash[17031]: cluster 2026-03-07T07:00:26.691006+0000 mgr.vm07.yrfcuj (mgr.14201) 324 : cluster [DBG] pgmap v202: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 511 B/s rd, 1023 B/s wr, 1 op/s 2026-03-07T08:00:30.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:29 vm09 bash[21220]: cluster 2026-03-07T07:00:28.691489+0000 mgr.vm07.yrfcuj (mgr.14201) 325 : cluster [DBG] pgmap v203: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:29 vm09 bash[21220]: cluster 2026-03-07T07:00:28.691489+0000 mgr.vm07.yrfcuj (mgr.14201) 325 : cluster [DBG] pgmap v203: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:30.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:29 vm07 bash[17031]: cluster 2026-03-07T07:00:28.691489+0000 mgr.vm07.yrfcuj (mgr.14201) 325 : cluster [DBG] pgmap v203: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:30.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:29 vm07 bash[17031]: cluster 2026-03-07T07:00:28.691489+0000 mgr.vm07.yrfcuj (mgr.14201) 325 : cluster [DBG] pgmap v203: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:31.580 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:00:31.744 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:00:31.745 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 41s ago 2m - - 2026-03-07T08:00:31.745 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 41s ago 2m 100M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:00:31.745 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 41s ago 2m 95.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:00:31.745 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (2m) 41s ago 2m 95.8M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:00:31.932 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:00:31.932 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:00:31.932 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:00:32.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:31 vm09 bash[21220]: cluster 2026-03-07T07:00:30.691835+0000 mgr.vm07.yrfcuj (mgr.14201) 326 : cluster [DBG] pgmap v204: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:32.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:31 vm09 bash[21220]: cluster 2026-03-07T07:00:30.691835+0000 mgr.vm07.yrfcuj (mgr.14201) 326 : cluster [DBG] pgmap v204: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:32.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:31 vm07 bash[17031]: cluster 2026-03-07T07:00:30.691835+0000 mgr.vm07.yrfcuj (mgr.14201) 326 : cluster [DBG] pgmap v204: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:32.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:31 vm07 bash[17031]: cluster 2026-03-07T07:00:30.691835+0000 mgr.vm07.yrfcuj (mgr.14201) 326 : cluster [DBG] pgmap v204: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:33.079 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:32 vm09 bash[21220]: audit 2026-03-07T07:00:31.569686+0000 mgr.vm07.yrfcuj (mgr.14201) 327 : audit [DBG] from='client.14916 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:33.079 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:32 vm09 bash[21220]: audit 2026-03-07T07:00:31.569686+0000 mgr.vm07.yrfcuj (mgr.14201) 327 : audit [DBG] from='client.14916 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:33.080 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:32 vm09 bash[21220]: audit 2026-03-07T07:00:31.747094+0000 mgr.vm07.yrfcuj (mgr.14201) 328 : audit [DBG] from='client.14920 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:33.080 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:32 vm09 bash[21220]: audit 2026-03-07T07:00:31.747094+0000 mgr.vm07.yrfcuj (mgr.14201) 328 : audit [DBG] from='client.14920 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:33.080 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:32 vm09 bash[21220]: audit 2026-03-07T07:00:31.936907+0000 mon.vm07 (mon.0) 862 : audit [DBG] from='client.? 192.168.123.107:0/674845026' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:33.080 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:32 vm09 bash[21220]: audit 2026-03-07T07:00:31.936907+0000 mon.vm07 (mon.0) 862 : audit [DBG] from='client.? 192.168.123.107:0/674845026' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:33.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:32 vm07 bash[17031]: audit 2026-03-07T07:00:31.569686+0000 mgr.vm07.yrfcuj (mgr.14201) 327 : audit [DBG] from='client.14916 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:33.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:32 vm07 bash[17031]: audit 2026-03-07T07:00:31.569686+0000 mgr.vm07.yrfcuj (mgr.14201) 327 : audit [DBG] from='client.14916 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:33.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:32 vm07 bash[17031]: audit 2026-03-07T07:00:31.747094+0000 mgr.vm07.yrfcuj (mgr.14201) 328 : audit [DBG] from='client.14920 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:33.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:32 vm07 bash[17031]: audit 2026-03-07T07:00:31.747094+0000 mgr.vm07.yrfcuj (mgr.14201) 328 : audit [DBG] from='client.14920 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:33.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:32 vm07 bash[17031]: audit 2026-03-07T07:00:31.936907+0000 mon.vm07 (mon.0) 862 : audit [DBG] from='client.? 192.168.123.107:0/674845026' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:33.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:32 vm07 bash[17031]: audit 2026-03-07T07:00:31.936907+0000 mon.vm07 (mon.0) 862 : audit [DBG] from='client.? 192.168.123.107:0/674845026' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:34.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:33 vm09 bash[21220]: cluster 2026-03-07T07:00:32.692220+0000 mgr.vm07.yrfcuj (mgr.14201) 329 : cluster [DBG] pgmap v205: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:34.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:33 vm09 bash[21220]: cluster 2026-03-07T07:00:32.692220+0000 mgr.vm07.yrfcuj (mgr.14201) 329 : cluster [DBG] pgmap v205: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:34.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:33 vm07 bash[17031]: cluster 2026-03-07T07:00:32.692220+0000 mgr.vm07.yrfcuj (mgr.14201) 329 : cluster [DBG] pgmap v205: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:34.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:33 vm07 bash[17031]: cluster 2026-03-07T07:00:32.692220+0000 mgr.vm07.yrfcuj (mgr.14201) 329 : cluster [DBG] pgmap v205: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:35 vm09 bash[21220]: cluster 2026-03-07T07:00:34.692567+0000 mgr.vm07.yrfcuj (mgr.14201) 330 : cluster [DBG] pgmap v206: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:35 vm09 bash[21220]: cluster 2026-03-07T07:00:34.692567+0000 mgr.vm07.yrfcuj (mgr.14201) 330 : cluster [DBG] pgmap v206: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:36.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:35 vm07 bash[17031]: cluster 2026-03-07T07:00:34.692567+0000 mgr.vm07.yrfcuj (mgr.14201) 330 : cluster [DBG] pgmap v206: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:36.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:35 vm07 bash[17031]: cluster 2026-03-07T07:00:34.692567+0000 mgr.vm07.yrfcuj (mgr.14201) 330 : cluster [DBG] pgmap v206: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:37.115 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:00:37.278 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:00:37.278 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 46s ago 2m - - 2026-03-07T08:00:37.278 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 46s ago 2m 100M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:00:37.278 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 47s ago 3m 95.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:00:37.278 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (2m) 47s ago 2m 95.8M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:00:37.469 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:00:37.469 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:00:37.469 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:00:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:37 vm09 bash[21220]: cluster 2026-03-07T07:00:36.692931+0000 mgr.vm07.yrfcuj (mgr.14201) 331 : cluster [DBG] pgmap v207: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:37 vm09 bash[21220]: cluster 2026-03-07T07:00:36.692931+0000 mgr.vm07.yrfcuj (mgr.14201) 331 : cluster [DBG] pgmap v207: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:37 vm09 bash[21220]: audit 2026-03-07T07:00:37.105048+0000 mgr.vm07.yrfcuj (mgr.14201) 332 : audit [DBG] from='client.14928 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:37 vm09 bash[21220]: audit 2026-03-07T07:00:37.105048+0000 mgr.vm07.yrfcuj (mgr.14201) 332 : audit [DBG] from='client.14928 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:37 vm09 bash[21220]: audit 2026-03-07T07:00:37.279621+0000 mgr.vm07.yrfcuj (mgr.14201) 333 : audit [DBG] from='client.14932 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:37 vm09 bash[21220]: audit 2026-03-07T07:00:37.279621+0000 mgr.vm07.yrfcuj (mgr.14201) 333 : audit [DBG] from='client.14932 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:37 vm09 bash[21220]: audit 2026-03-07T07:00:37.473760+0000 mon.vm07 (mon.0) 863 : audit [DBG] from='client.? 192.168.123.107:0/2803202999' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:37 vm09 bash[21220]: audit 2026-03-07T07:00:37.473760+0000 mon.vm07 (mon.0) 863 : audit [DBG] from='client.? 192.168.123.107:0/2803202999' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:38.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:37 vm07 bash[17031]: cluster 2026-03-07T07:00:36.692931+0000 mgr.vm07.yrfcuj (mgr.14201) 331 : cluster [DBG] pgmap v207: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:38.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:37 vm07 bash[17031]: cluster 2026-03-07T07:00:36.692931+0000 mgr.vm07.yrfcuj (mgr.14201) 331 : cluster [DBG] pgmap v207: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:38.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:37 vm07 bash[17031]: audit 2026-03-07T07:00:37.105048+0000 mgr.vm07.yrfcuj (mgr.14201) 332 : audit [DBG] from='client.14928 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:38.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:37 vm07 bash[17031]: audit 2026-03-07T07:00:37.105048+0000 mgr.vm07.yrfcuj (mgr.14201) 332 : audit [DBG] from='client.14928 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:38.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:37 vm07 bash[17031]: audit 2026-03-07T07:00:37.279621+0000 mgr.vm07.yrfcuj (mgr.14201) 333 : audit [DBG] from='client.14932 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:38.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:37 vm07 bash[17031]: audit 2026-03-07T07:00:37.279621+0000 mgr.vm07.yrfcuj (mgr.14201) 333 : audit [DBG] from='client.14932 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:38.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:37 vm07 bash[17031]: audit 2026-03-07T07:00:37.473760+0000 mon.vm07 (mon.0) 863 : audit [DBG] from='client.? 192.168.123.107:0/2803202999' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:38.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:37 vm07 bash[17031]: audit 2026-03-07T07:00:37.473760+0000 mon.vm07 (mon.0) 863 : audit [DBG] from='client.? 192.168.123.107:0/2803202999' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:40.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:39 vm09 bash[21220]: cluster 2026-03-07T07:00:38.693420+0000 mgr.vm07.yrfcuj (mgr.14201) 334 : cluster [DBG] pgmap v208: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:40.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:39 vm09 bash[21220]: cluster 2026-03-07T07:00:38.693420+0000 mgr.vm07.yrfcuj (mgr.14201) 334 : cluster [DBG] pgmap v208: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:40.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:40 vm07 bash[17031]: cluster 2026-03-07T07:00:38.693420+0000 mgr.vm07.yrfcuj (mgr.14201) 334 : cluster [DBG] pgmap v208: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:40.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:40 vm07 bash[17031]: cluster 2026-03-07T07:00:38.693420+0000 mgr.vm07.yrfcuj (mgr.14201) 334 : cluster [DBG] pgmap v208: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:41.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:41 vm09 bash[21220]: audit 2026-03-07T07:00:40.130299+0000 mon.vm07 (mon.0) 864 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:00:41.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:41 vm09 bash[21220]: audit 2026-03-07T07:00:40.130299+0000 mon.vm07 (mon.0) 864 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:00:41.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:41 vm07 bash[17031]: audit 2026-03-07T07:00:40.130299+0000 mon.vm07 (mon.0) 864 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:00:41.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:41 vm07 bash[17031]: audit 2026-03-07T07:00:40.130299+0000 mon.vm07 (mon.0) 864 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:00:42.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:42 vm09 bash[21220]: cluster 2026-03-07T07:00:40.693822+0000 mgr.vm07.yrfcuj (mgr.14201) 335 : cluster [DBG] pgmap v209: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:42.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:42 vm09 bash[21220]: cluster 2026-03-07T07:00:40.693822+0000 mgr.vm07.yrfcuj (mgr.14201) 335 : cluster [DBG] pgmap v209: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:42.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:42 vm07 bash[17031]: cluster 2026-03-07T07:00:40.693822+0000 mgr.vm07.yrfcuj (mgr.14201) 335 : cluster [DBG] pgmap v209: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:42.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:42 vm07 bash[17031]: cluster 2026-03-07T07:00:40.693822+0000 mgr.vm07.yrfcuj (mgr.14201) 335 : cluster [DBG] pgmap v209: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:42.651 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:00:42.813 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:00:42.813 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 52s ago 3m - - 2026-03-07T08:00:42.813 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 52s ago 3m 100M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:00:42.813 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 52s ago 3m 95.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:00:42.813 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (3m) 52s ago 3m 95.8M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:00:43.006 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:00:43.006 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:00:43.006 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:00:43.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:43 vm09 bash[21220]: audit 2026-03-07T07:00:43.010437+0000 mon.vm07 (mon.0) 865 : audit [DBG] from='client.? 192.168.123.107:0/3125170769' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:43.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:43 vm09 bash[21220]: audit 2026-03-07T07:00:43.010437+0000 mon.vm07 (mon.0) 865 : audit [DBG] from='client.? 192.168.123.107:0/3125170769' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:43.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:43 vm07 bash[17031]: audit 2026-03-07T07:00:43.010437+0000 mon.vm07 (mon.0) 865 : audit [DBG] from='client.? 192.168.123.107:0/3125170769' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:43.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:43 vm07 bash[17031]: audit 2026-03-07T07:00:43.010437+0000 mon.vm07 (mon.0) 865 : audit [DBG] from='client.? 192.168.123.107:0/3125170769' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:44.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:44 vm09 bash[21220]: audit 2026-03-07T07:00:42.639225+0000 mgr.vm07.yrfcuj (mgr.14201) 336 : audit [DBG] from='client.14940 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:44.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:44 vm09 bash[21220]: audit 2026-03-07T07:00:42.639225+0000 mgr.vm07.yrfcuj (mgr.14201) 336 : audit [DBG] from='client.14940 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:44.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:44 vm09 bash[21220]: cluster 2026-03-07T07:00:42.694342+0000 mgr.vm07.yrfcuj (mgr.14201) 337 : cluster [DBG] pgmap v210: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:44.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:44 vm09 bash[21220]: cluster 2026-03-07T07:00:42.694342+0000 mgr.vm07.yrfcuj (mgr.14201) 337 : cluster [DBG] pgmap v210: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:44.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:44 vm09 bash[21220]: audit 2026-03-07T07:00:42.814488+0000 mgr.vm07.yrfcuj (mgr.14201) 338 : audit [DBG] from='client.14944 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:44.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:44 vm09 bash[21220]: audit 2026-03-07T07:00:42.814488+0000 mgr.vm07.yrfcuj (mgr.14201) 338 : audit [DBG] from='client.14944 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:44.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:44 vm07 bash[17031]: audit 2026-03-07T07:00:42.639225+0000 mgr.vm07.yrfcuj (mgr.14201) 336 : audit [DBG] from='client.14940 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:44.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:44 vm07 bash[17031]: audit 2026-03-07T07:00:42.639225+0000 mgr.vm07.yrfcuj (mgr.14201) 336 : audit [DBG] from='client.14940 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:44.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:44 vm07 bash[17031]: cluster 2026-03-07T07:00:42.694342+0000 mgr.vm07.yrfcuj (mgr.14201) 337 : cluster [DBG] pgmap v210: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:44.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:44 vm07 bash[17031]: cluster 2026-03-07T07:00:42.694342+0000 mgr.vm07.yrfcuj (mgr.14201) 337 : cluster [DBG] pgmap v210: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:44.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:44 vm07 bash[17031]: audit 2026-03-07T07:00:42.814488+0000 mgr.vm07.yrfcuj (mgr.14201) 338 : audit [DBG] from='client.14944 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:44.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:44 vm07 bash[17031]: audit 2026-03-07T07:00:42.814488+0000 mgr.vm07.yrfcuj (mgr.14201) 338 : audit [DBG] from='client.14944 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:46.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:46 vm09 bash[21220]: cluster 2026-03-07T07:00:44.694756+0000 mgr.vm07.yrfcuj (mgr.14201) 339 : cluster [DBG] pgmap v211: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:46.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:46 vm09 bash[21220]: cluster 2026-03-07T07:00:44.694756+0000 mgr.vm07.yrfcuj (mgr.14201) 339 : cluster [DBG] pgmap v211: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:46.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:46 vm07 bash[17031]: cluster 2026-03-07T07:00:44.694756+0000 mgr.vm07.yrfcuj (mgr.14201) 339 : cluster [DBG] pgmap v211: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:46.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:46 vm07 bash[17031]: cluster 2026-03-07T07:00:44.694756+0000 mgr.vm07.yrfcuj (mgr.14201) 339 : cluster [DBG] pgmap v211: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:48.185 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:00:48.333 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:00:48.333 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 57s ago 3m - - 2026-03-07T08:00:48.333 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 57s ago 3m 100M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:00:48.333 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 58s ago 3m 95.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:00:48.333 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (3m) 58s ago 3m 95.8M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:00:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:48 vm09 bash[21220]: cluster 2026-03-07T07:00:46.695190+0000 mgr.vm07.yrfcuj (mgr.14201) 340 : cluster [DBG] pgmap v212: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:48 vm09 bash[21220]: cluster 2026-03-07T07:00:46.695190+0000 mgr.vm07.yrfcuj (mgr.14201) 340 : cluster [DBG] pgmap v212: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:48.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:48 vm07 bash[17031]: cluster 2026-03-07T07:00:46.695190+0000 mgr.vm07.yrfcuj (mgr.14201) 340 : cluster [DBG] pgmap v212: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:48.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:48 vm07 bash[17031]: cluster 2026-03-07T07:00:46.695190+0000 mgr.vm07.yrfcuj (mgr.14201) 340 : cluster [DBG] pgmap v212: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:48.520 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:00:48.520 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:00:48.520 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:00:49.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:49 vm07 bash[17031]: audit 2026-03-07T07:00:48.174168+0000 mgr.vm07.yrfcuj (mgr.14201) 341 : audit [DBG] from='client.14952 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:49.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:49 vm07 bash[17031]: audit 2026-03-07T07:00:48.174168+0000 mgr.vm07.yrfcuj (mgr.14201) 341 : audit [DBG] from='client.14952 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:49.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:49 vm07 bash[17031]: audit 2026-03-07T07:00:48.334670+0000 mgr.vm07.yrfcuj (mgr.14201) 342 : audit [DBG] from='client.14956 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:49.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:49 vm07 bash[17031]: audit 2026-03-07T07:00:48.334670+0000 mgr.vm07.yrfcuj (mgr.14201) 342 : audit [DBG] from='client.14956 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:49.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:49 vm07 bash[17031]: audit 2026-03-07T07:00:48.524483+0000 mon.vm07 (mon.0) 866 : audit [DBG] from='client.? 192.168.123.107:0/4187735759' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:49.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:49 vm07 bash[17031]: audit 2026-03-07T07:00:48.524483+0000 mon.vm07 (mon.0) 866 : audit [DBG] from='client.? 192.168.123.107:0/4187735759' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:49.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:49 vm09 bash[21220]: audit 2026-03-07T07:00:48.174168+0000 mgr.vm07.yrfcuj (mgr.14201) 341 : audit [DBG] from='client.14952 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:49.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:49 vm09 bash[21220]: audit 2026-03-07T07:00:48.174168+0000 mgr.vm07.yrfcuj (mgr.14201) 341 : audit [DBG] from='client.14952 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:49.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:49 vm09 bash[21220]: audit 2026-03-07T07:00:48.334670+0000 mgr.vm07.yrfcuj (mgr.14201) 342 : audit [DBG] from='client.14956 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:49.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:49 vm09 bash[21220]: audit 2026-03-07T07:00:48.334670+0000 mgr.vm07.yrfcuj (mgr.14201) 342 : audit [DBG] from='client.14956 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:49.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:49 vm09 bash[21220]: audit 2026-03-07T07:00:48.524483+0000 mon.vm07 (mon.0) 866 : audit [DBG] from='client.? 192.168.123.107:0/4187735759' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:49.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:49 vm09 bash[21220]: audit 2026-03-07T07:00:48.524483+0000 mon.vm07 (mon.0) 866 : audit [DBG] from='client.? 192.168.123.107:0/4187735759' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:50.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:50 vm07 bash[17031]: cluster 2026-03-07T07:00:48.695630+0000 mgr.vm07.yrfcuj (mgr.14201) 343 : cluster [DBG] pgmap v213: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:50.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:50 vm07 bash[17031]: cluster 2026-03-07T07:00:48.695630+0000 mgr.vm07.yrfcuj (mgr.14201) 343 : cluster [DBG] pgmap v213: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:50.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:50 vm09 bash[21220]: cluster 2026-03-07T07:00:48.695630+0000 mgr.vm07.yrfcuj (mgr.14201) 343 : cluster [DBG] pgmap v213: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:50.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:50 vm09 bash[21220]: cluster 2026-03-07T07:00:48.695630+0000 mgr.vm07.yrfcuj (mgr.14201) 343 : cluster [DBG] pgmap v213: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:51.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:51 vm07 bash[17031]: audit 2026-03-07T07:00:50.733823+0000 mon.vm07 (mon.0) 867 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:00:51.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:51 vm07 bash[17031]: audit 2026-03-07T07:00:50.733823+0000 mon.vm07 (mon.0) 867 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:00:51.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:51 vm09 bash[21220]: audit 2026-03-07T07:00:50.733823+0000 mon.vm07 (mon.0) 867 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:00:51.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:51 vm09 bash[21220]: audit 2026-03-07T07:00:50.733823+0000 mon.vm07 (mon.0) 867 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:00:52.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:52 vm07 bash[17031]: cluster 2026-03-07T07:00:50.696041+0000 mgr.vm07.yrfcuj (mgr.14201) 344 : cluster [DBG] pgmap v214: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:52.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:52 vm07 bash[17031]: cluster 2026-03-07T07:00:50.696041+0000 mgr.vm07.yrfcuj (mgr.14201) 344 : cluster [DBG] pgmap v214: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:52.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:52 vm09 bash[21220]: cluster 2026-03-07T07:00:50.696041+0000 mgr.vm07.yrfcuj (mgr.14201) 344 : cluster [DBG] pgmap v214: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:52.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:52 vm09 bash[21220]: cluster 2026-03-07T07:00:50.696041+0000 mgr.vm07.yrfcuj (mgr.14201) 344 : cluster [DBG] pgmap v214: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:53.696 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:00:53.843 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:00:53.843 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 63s ago 3m - - 2026-03-07T08:00:53.843 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 63s ago 3m 100M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:00:53.843 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 63s ago 3m 95.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:00:53.843 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (3m) 63s ago 3m 95.8M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:00:54.060 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:00:54.060 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:00:54.060 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:00:54.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:54 vm09 bash[21220]: cluster 2026-03-07T07:00:52.696511+0000 mgr.vm07.yrfcuj (mgr.14201) 345 : cluster [DBG] pgmap v215: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:54.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:54 vm09 bash[21220]: cluster 2026-03-07T07:00:52.696511+0000 mgr.vm07.yrfcuj (mgr.14201) 345 : cluster [DBG] pgmap v215: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:54.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:54 vm09 bash[21220]: audit 2026-03-07T07:00:54.064698+0000 mon.vm07 (mon.0) 868 : audit [DBG] from='client.? 192.168.123.107:0/154422886' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:54.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:54 vm09 bash[21220]: audit 2026-03-07T07:00:54.064698+0000 mon.vm07 (mon.0) 868 : audit [DBG] from='client.? 192.168.123.107:0/154422886' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:54.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:54 vm07 bash[17031]: cluster 2026-03-07T07:00:52.696511+0000 mgr.vm07.yrfcuj (mgr.14201) 345 : cluster [DBG] pgmap v215: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:54.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:54 vm07 bash[17031]: cluster 2026-03-07T07:00:52.696511+0000 mgr.vm07.yrfcuj (mgr.14201) 345 : cluster [DBG] pgmap v215: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:54.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:54 vm07 bash[17031]: audit 2026-03-07T07:00:54.064698+0000 mon.vm07 (mon.0) 868 : audit [DBG] from='client.? 192.168.123.107:0/154422886' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:54.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:54 vm07 bash[17031]: audit 2026-03-07T07:00:54.064698+0000 mon.vm07 (mon.0) 868 : audit [DBG] from='client.? 192.168.123.107:0/154422886' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:00:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:55 vm09 bash[21220]: audit 2026-03-07T07:00:53.685793+0000 mgr.vm07.yrfcuj (mgr.14201) 346 : audit [DBG] from='client.14964 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:55 vm09 bash[21220]: audit 2026-03-07T07:00:53.685793+0000 mgr.vm07.yrfcuj (mgr.14201) 346 : audit [DBG] from='client.14964 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:55 vm09 bash[21220]: audit 2026-03-07T07:00:53.845220+0000 mgr.vm07.yrfcuj (mgr.14201) 347 : audit [DBG] from='client.14968 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:55 vm09 bash[21220]: audit 2026-03-07T07:00:53.845220+0000 mgr.vm07.yrfcuj (mgr.14201) 347 : audit [DBG] from='client.14968 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:55 vm09 bash[21220]: audit 2026-03-07T07:00:55.130639+0000 mon.vm07 (mon.0) 869 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:00:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:55 vm09 bash[21220]: audit 2026-03-07T07:00:55.130639+0000 mon.vm07 (mon.0) 869 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:00:55.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:55 vm07 bash[17031]: audit 2026-03-07T07:00:53.685793+0000 mgr.vm07.yrfcuj (mgr.14201) 346 : audit [DBG] from='client.14964 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:55.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:55 vm07 bash[17031]: audit 2026-03-07T07:00:53.685793+0000 mgr.vm07.yrfcuj (mgr.14201) 346 : audit [DBG] from='client.14964 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:55.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:55 vm07 bash[17031]: audit 2026-03-07T07:00:53.845220+0000 mgr.vm07.yrfcuj (mgr.14201) 347 : audit [DBG] from='client.14968 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:55.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:55 vm07 bash[17031]: audit 2026-03-07T07:00:53.845220+0000 mgr.vm07.yrfcuj (mgr.14201) 347 : audit [DBG] from='client.14968 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:00:55.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:55 vm07 bash[17031]: audit 2026-03-07T07:00:55.130639+0000 mon.vm07 (mon.0) 869 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:00:55.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:55 vm07 bash[17031]: audit 2026-03-07T07:00:55.130639+0000 mon.vm07 (mon.0) 869 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:00:56.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:56 vm09 bash[21220]: cluster 2026-03-07T07:00:54.696921+0000 mgr.vm07.yrfcuj (mgr.14201) 348 : cluster [DBG] pgmap v216: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:56 vm09 bash[21220]: cluster 2026-03-07T07:00:54.696921+0000 mgr.vm07.yrfcuj (mgr.14201) 348 : cluster [DBG] pgmap v216: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:56 vm09 bash[21220]: audit 2026-03-07T07:00:56.044966+0000 mon.vm07 (mon.0) 870 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:56 vm09 bash[21220]: audit 2026-03-07T07:00:56.044966+0000 mon.vm07 (mon.0) 870 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:56 vm09 bash[21220]: audit 2026-03-07T07:00:56.050380+0000 mon.vm07 (mon.0) 871 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:56 vm09 bash[21220]: audit 2026-03-07T07:00:56.050380+0000 mon.vm07 (mon.0) 871 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:56.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:56 vm07 bash[17031]: cluster 2026-03-07T07:00:54.696921+0000 mgr.vm07.yrfcuj (mgr.14201) 348 : cluster [DBG] pgmap v216: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:56.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:56 vm07 bash[17031]: cluster 2026-03-07T07:00:54.696921+0000 mgr.vm07.yrfcuj (mgr.14201) 348 : cluster [DBG] pgmap v216: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:00:56.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:56 vm07 bash[17031]: audit 2026-03-07T07:00:56.044966+0000 mon.vm07 (mon.0) 870 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:56.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:56 vm07 bash[17031]: audit 2026-03-07T07:00:56.044966+0000 mon.vm07 (mon.0) 870 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:56.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:56 vm07 bash[17031]: audit 2026-03-07T07:00:56.050380+0000 mon.vm07 (mon.0) 871 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:56.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:56 vm07 bash[17031]: audit 2026-03-07T07:00:56.050380+0000 mon.vm07 (mon.0) 871 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:57 vm09 bash[21220]: audit 2026-03-07T07:00:56.293268+0000 mon.vm07 (mon.0) 872 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:57 vm09 bash[21220]: audit 2026-03-07T07:00:56.293268+0000 mon.vm07 (mon.0) 872 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:57 vm09 bash[21220]: audit 2026-03-07T07:00:56.298342+0000 mon.vm07 (mon.0) 873 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:57 vm09 bash[21220]: audit 2026-03-07T07:00:56.298342+0000 mon.vm07 (mon.0) 873 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:57 vm09 bash[21220]: audit 2026-03-07T07:00:56.591682+0000 mon.vm07 (mon.0) 874 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:00:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:57 vm09 bash[21220]: audit 2026-03-07T07:00:56.591682+0000 mon.vm07 (mon.0) 874 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:00:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:57 vm09 bash[21220]: audit 2026-03-07T07:00:56.592167+0000 mon.vm07 (mon.0) 875 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:00:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:57 vm09 bash[21220]: audit 2026-03-07T07:00:56.592167+0000 mon.vm07 (mon.0) 875 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:00:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:57 vm09 bash[21220]: audit 2026-03-07T07:00:56.597140+0000 mon.vm07 (mon.0) 876 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:57 vm09 bash[21220]: audit 2026-03-07T07:00:56.597140+0000 mon.vm07 (mon.0) 876 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:57 vm09 bash[21220]: audit 2026-03-07T07:00:56.598397+0000 mon.vm07 (mon.0) 877 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:00:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:57 vm09 bash[21220]: audit 2026-03-07T07:00:56.598397+0000 mon.vm07 (mon.0) 877 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:00:57.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:57 vm07 bash[17031]: audit 2026-03-07T07:00:56.293268+0000 mon.vm07 (mon.0) 872 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:57.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:57 vm07 bash[17031]: audit 2026-03-07T07:00:56.293268+0000 mon.vm07 (mon.0) 872 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:57.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:57 vm07 bash[17031]: audit 2026-03-07T07:00:56.298342+0000 mon.vm07 (mon.0) 873 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:57.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:57 vm07 bash[17031]: audit 2026-03-07T07:00:56.298342+0000 mon.vm07 (mon.0) 873 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:57.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:57 vm07 bash[17031]: audit 2026-03-07T07:00:56.591682+0000 mon.vm07 (mon.0) 874 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:00:57.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:57 vm07 bash[17031]: audit 2026-03-07T07:00:56.591682+0000 mon.vm07 (mon.0) 874 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:00:57.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:57 vm07 bash[17031]: audit 2026-03-07T07:00:56.592167+0000 mon.vm07 (mon.0) 875 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:00:57.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:57 vm07 bash[17031]: audit 2026-03-07T07:00:56.592167+0000 mon.vm07 (mon.0) 875 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:00:57.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:57 vm07 bash[17031]: audit 2026-03-07T07:00:56.597140+0000 mon.vm07 (mon.0) 876 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:57.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:57 vm07 bash[17031]: audit 2026-03-07T07:00:56.597140+0000 mon.vm07 (mon.0) 876 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:00:57.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:57 vm07 bash[17031]: audit 2026-03-07T07:00:56.598397+0000 mon.vm07 (mon.0) 877 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:00:57.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:57 vm07 bash[17031]: audit 2026-03-07T07:00:56.598397+0000 mon.vm07 (mon.0) 877 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:00:58.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:58 vm09 bash[21220]: cluster 2026-03-07T07:00:56.592964+0000 mgr.vm07.yrfcuj (mgr.14201) 349 : cluster [DBG] pgmap v217: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 258 B/s rd, 516 B/s wr, 0 op/s 2026-03-07T08:00:58.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:58 vm09 bash[21220]: cluster 2026-03-07T07:00:56.592964+0000 mgr.vm07.yrfcuj (mgr.14201) 349 : cluster [DBG] pgmap v217: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 258 B/s rd, 516 B/s wr, 0 op/s 2026-03-07T08:00:58.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:58 vm09 bash[21220]: cluster 2026-03-07T07:00:56.593079+0000 mgr.vm07.yrfcuj (mgr.14201) 350 : cluster [DBG] pgmap v218: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:58.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:00:58 vm09 bash[21220]: cluster 2026-03-07T07:00:56.593079+0000 mgr.vm07.yrfcuj (mgr.14201) 350 : cluster [DBG] pgmap v218: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:58.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:58 vm07 bash[17031]: cluster 2026-03-07T07:00:56.592964+0000 mgr.vm07.yrfcuj (mgr.14201) 349 : cluster [DBG] pgmap v217: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 258 B/s rd, 516 B/s wr, 0 op/s 2026-03-07T08:00:58.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:58 vm07 bash[17031]: cluster 2026-03-07T07:00:56.592964+0000 mgr.vm07.yrfcuj (mgr.14201) 349 : cluster [DBG] pgmap v217: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 258 B/s rd, 516 B/s wr, 0 op/s 2026-03-07T08:00:58.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:58 vm07 bash[17031]: cluster 2026-03-07T07:00:56.593079+0000 mgr.vm07.yrfcuj (mgr.14201) 350 : cluster [DBG] pgmap v218: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:58.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:00:58 vm07 bash[17031]: cluster 2026-03-07T07:00:56.593079+0000 mgr.vm07.yrfcuj (mgr.14201) 350 : cluster [DBG] pgmap v218: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:00:59.246 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:00:59.403 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:00:59.403 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 3s ago 3m - - 2026-03-07T08:00:59.403 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 3s ago 3m 105M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:00:59.403 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 3s ago 3m 98.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:00:59.403 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (3m) 3s ago 3m 98.9M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:00:59.604 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:00:59.604 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:00:59.604 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:01:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:00 vm09 bash[21220]: cluster 2026-03-07T07:00:58.593513+0000 mgr.vm07.yrfcuj (mgr.14201) 351 : cluster [DBG] pgmap v219: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:00 vm09 bash[21220]: cluster 2026-03-07T07:00:58.593513+0000 mgr.vm07.yrfcuj (mgr.14201) 351 : cluster [DBG] pgmap v219: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:00 vm09 bash[21220]: audit 2026-03-07T07:00:59.233575+0000 mgr.vm07.yrfcuj (mgr.14201) 352 : audit [DBG] from='client.14976 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:00 vm09 bash[21220]: audit 2026-03-07T07:00:59.233575+0000 mgr.vm07.yrfcuj (mgr.14201) 352 : audit [DBG] from='client.14976 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:00 vm09 bash[21220]: audit 2026-03-07T07:00:59.404226+0000 mgr.vm07.yrfcuj (mgr.14201) 353 : audit [DBG] from='client.14980 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:00 vm09 bash[21220]: audit 2026-03-07T07:00:59.404226+0000 mgr.vm07.yrfcuj (mgr.14201) 353 : audit [DBG] from='client.14980 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:00 vm09 bash[21220]: audit 2026-03-07T07:00:59.608554+0000 mon.vm07 (mon.0) 878 : audit [DBG] from='client.? 192.168.123.107:0/2818605487' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:00 vm09 bash[21220]: audit 2026-03-07T07:00:59.608554+0000 mon.vm07 (mon.0) 878 : audit [DBG] from='client.? 192.168.123.107:0/2818605487' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:00.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:00 vm07 bash[17031]: cluster 2026-03-07T07:00:58.593513+0000 mgr.vm07.yrfcuj (mgr.14201) 351 : cluster [DBG] pgmap v219: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:00.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:00 vm07 bash[17031]: cluster 2026-03-07T07:00:58.593513+0000 mgr.vm07.yrfcuj (mgr.14201) 351 : cluster [DBG] pgmap v219: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:00.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:00 vm07 bash[17031]: audit 2026-03-07T07:00:59.233575+0000 mgr.vm07.yrfcuj (mgr.14201) 352 : audit [DBG] from='client.14976 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:00.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:00 vm07 bash[17031]: audit 2026-03-07T07:00:59.233575+0000 mgr.vm07.yrfcuj (mgr.14201) 352 : audit [DBG] from='client.14976 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:00.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:00 vm07 bash[17031]: audit 2026-03-07T07:00:59.404226+0000 mgr.vm07.yrfcuj (mgr.14201) 353 : audit [DBG] from='client.14980 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:00.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:00 vm07 bash[17031]: audit 2026-03-07T07:00:59.404226+0000 mgr.vm07.yrfcuj (mgr.14201) 353 : audit [DBG] from='client.14980 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:00.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:00 vm07 bash[17031]: audit 2026-03-07T07:00:59.608554+0000 mon.vm07 (mon.0) 878 : audit [DBG] from='client.? 192.168.123.107:0/2818605487' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:00.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:00 vm07 bash[17031]: audit 2026-03-07T07:00:59.608554+0000 mon.vm07 (mon.0) 878 : audit [DBG] from='client.? 192.168.123.107:0/2818605487' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:01.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:01 vm07 bash[17031]: cluster 2026-03-07T07:01:00.593912+0000 mgr.vm07.yrfcuj (mgr.14201) 354 : cluster [DBG] pgmap v220: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:01.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:01 vm07 bash[17031]: cluster 2026-03-07T07:01:00.593912+0000 mgr.vm07.yrfcuj (mgr.14201) 354 : cluster [DBG] pgmap v220: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:02.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:01 vm09 bash[21220]: cluster 2026-03-07T07:01:00.593912+0000 mgr.vm07.yrfcuj (mgr.14201) 354 : cluster [DBG] pgmap v220: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:02.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:01 vm09 bash[21220]: cluster 2026-03-07T07:01:00.593912+0000 mgr.vm07.yrfcuj (mgr.14201) 354 : cluster [DBG] pgmap v220: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:03.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:03 vm07 bash[17031]: cluster 2026-03-07T07:01:02.594345+0000 mgr.vm07.yrfcuj (mgr.14201) 355 : cluster [DBG] pgmap v221: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:03.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:03 vm07 bash[17031]: cluster 2026-03-07T07:01:02.594345+0000 mgr.vm07.yrfcuj (mgr.14201) 355 : cluster [DBG] pgmap v221: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:04.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:03 vm09 bash[21220]: cluster 2026-03-07T07:01:02.594345+0000 mgr.vm07.yrfcuj (mgr.14201) 355 : cluster [DBG] pgmap v221: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:04.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:03 vm09 bash[21220]: cluster 2026-03-07T07:01:02.594345+0000 mgr.vm07.yrfcuj (mgr.14201) 355 : cluster [DBG] pgmap v221: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:04.786 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:01:04.941 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:01:04.941 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 8s ago 3m - - 2026-03-07T08:01:04.941 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 8s ago 3m 105M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:01:04.941 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 8s ago 3m 98.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:01:04.941 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (3m) 8s ago 3m 98.9M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:01:05.136 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:01:05.136 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:01:05.136 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:01:05.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:05 vm07 bash[17031]: cluster 2026-03-07T07:01:04.594714+0000 mgr.vm07.yrfcuj (mgr.14201) 356 : cluster [DBG] pgmap v222: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:05.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:05 vm07 bash[17031]: cluster 2026-03-07T07:01:04.594714+0000 mgr.vm07.yrfcuj (mgr.14201) 356 : cluster [DBG] pgmap v222: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:05.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:05 vm07 bash[17031]: audit 2026-03-07T07:01:04.774046+0000 mgr.vm07.yrfcuj (mgr.14201) 357 : audit [DBG] from='client.14988 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:05.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:05 vm07 bash[17031]: audit 2026-03-07T07:01:04.774046+0000 mgr.vm07.yrfcuj (mgr.14201) 357 : audit [DBG] from='client.14988 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:05.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:05 vm07 bash[17031]: audit 2026-03-07T07:01:04.942590+0000 mgr.vm07.yrfcuj (mgr.14201) 358 : audit [DBG] from='client.14992 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:05.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:05 vm07 bash[17031]: audit 2026-03-07T07:01:04.942590+0000 mgr.vm07.yrfcuj (mgr.14201) 358 : audit [DBG] from='client.14992 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:05.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:05 vm07 bash[17031]: audit 2026-03-07T07:01:05.140410+0000 mon.vm07 (mon.0) 879 : audit [DBG] from='client.? 192.168.123.107:0/483538998' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:05.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:05 vm07 bash[17031]: audit 2026-03-07T07:01:05.140410+0000 mon.vm07 (mon.0) 879 : audit [DBG] from='client.? 192.168.123.107:0/483538998' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:06.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:05 vm09 bash[21220]: cluster 2026-03-07T07:01:04.594714+0000 mgr.vm07.yrfcuj (mgr.14201) 356 : cluster [DBG] pgmap v222: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:06.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:05 vm09 bash[21220]: cluster 2026-03-07T07:01:04.594714+0000 mgr.vm07.yrfcuj (mgr.14201) 356 : cluster [DBG] pgmap v222: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:06.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:05 vm09 bash[21220]: audit 2026-03-07T07:01:04.774046+0000 mgr.vm07.yrfcuj (mgr.14201) 357 : audit [DBG] from='client.14988 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:06.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:05 vm09 bash[21220]: audit 2026-03-07T07:01:04.774046+0000 mgr.vm07.yrfcuj (mgr.14201) 357 : audit [DBG] from='client.14988 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:06.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:05 vm09 bash[21220]: audit 2026-03-07T07:01:04.942590+0000 mgr.vm07.yrfcuj (mgr.14201) 358 : audit [DBG] from='client.14992 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:06.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:05 vm09 bash[21220]: audit 2026-03-07T07:01:04.942590+0000 mgr.vm07.yrfcuj (mgr.14201) 358 : audit [DBG] from='client.14992 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:06.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:05 vm09 bash[21220]: audit 2026-03-07T07:01:05.140410+0000 mon.vm07 (mon.0) 879 : audit [DBG] from='client.? 192.168.123.107:0/483538998' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:06.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:05 vm09 bash[21220]: audit 2026-03-07T07:01:05.140410+0000 mon.vm07 (mon.0) 879 : audit [DBG] from='client.? 192.168.123.107:0/483538998' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:08.081 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:07 vm09 bash[21220]: cluster 2026-03-07T07:01:06.595059+0000 mgr.vm07.yrfcuj (mgr.14201) 359 : cluster [DBG] pgmap v223: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:08.081 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:07 vm09 bash[21220]: cluster 2026-03-07T07:01:06.595059+0000 mgr.vm07.yrfcuj (mgr.14201) 359 : cluster [DBG] pgmap v223: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:08.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:07 vm07 bash[17031]: cluster 2026-03-07T07:01:06.595059+0000 mgr.vm07.yrfcuj (mgr.14201) 359 : cluster [DBG] pgmap v223: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:08.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:07 vm07 bash[17031]: cluster 2026-03-07T07:01:06.595059+0000 mgr.vm07.yrfcuj (mgr.14201) 359 : cluster [DBG] pgmap v223: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:10.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:09 vm09 bash[21220]: cluster 2026-03-07T07:01:08.595492+0000 mgr.vm07.yrfcuj (mgr.14201) 360 : cluster [DBG] pgmap v224: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:10.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:09 vm09 bash[21220]: cluster 2026-03-07T07:01:08.595492+0000 mgr.vm07.yrfcuj (mgr.14201) 360 : cluster [DBG] pgmap v224: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:10.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:09 vm07 bash[17031]: cluster 2026-03-07T07:01:08.595492+0000 mgr.vm07.yrfcuj (mgr.14201) 360 : cluster [DBG] pgmap v224: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:10.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:09 vm07 bash[17031]: cluster 2026-03-07T07:01:08.595492+0000 mgr.vm07.yrfcuj (mgr.14201) 360 : cluster [DBG] pgmap v224: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:10.318 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:01:10.472 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:01:10.472 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 14s ago 3m - - 2026-03-07T08:01:10.472 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 14s ago 3m 105M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:01:10.472 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 14s ago 3m 98.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:01:10.472 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (3m) 14s ago 3m 98.9M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:01:10.653 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:01:10.653 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:01:10.653 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:01:11.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:10 vm09 bash[21220]: audit 2026-03-07T07:01:10.130815+0000 mon.vm07 (mon.0) 880 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:01:11.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:10 vm09 bash[21220]: audit 2026-03-07T07:01:10.130815+0000 mon.vm07 (mon.0) 880 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:01:11.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:10 vm09 bash[21220]: audit 2026-03-07T07:01:10.306240+0000 mgr.vm07.yrfcuj (mgr.14201) 361 : audit [DBG] from='client.15000 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:11.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:10 vm09 bash[21220]: audit 2026-03-07T07:01:10.306240+0000 mgr.vm07.yrfcuj (mgr.14201) 361 : audit [DBG] from='client.15000 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:11.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:10 vm09 bash[21220]: audit 2026-03-07T07:01:10.473759+0000 mgr.vm07.yrfcuj (mgr.14201) 362 : audit [DBG] from='client.24641 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:11.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:10 vm09 bash[21220]: audit 2026-03-07T07:01:10.473759+0000 mgr.vm07.yrfcuj (mgr.14201) 362 : audit [DBG] from='client.24641 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:11.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:10 vm09 bash[21220]: audit 2026-03-07T07:01:10.657587+0000 mon.vm07 (mon.0) 881 : audit [DBG] from='client.? 192.168.123.107:0/1485722833' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:11.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:10 vm09 bash[21220]: audit 2026-03-07T07:01:10.657587+0000 mon.vm07 (mon.0) 881 : audit [DBG] from='client.? 192.168.123.107:0/1485722833' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:11.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:10 vm07 bash[17031]: audit 2026-03-07T07:01:10.130815+0000 mon.vm07 (mon.0) 880 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:01:11.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:10 vm07 bash[17031]: audit 2026-03-07T07:01:10.130815+0000 mon.vm07 (mon.0) 880 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:01:11.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:10 vm07 bash[17031]: audit 2026-03-07T07:01:10.306240+0000 mgr.vm07.yrfcuj (mgr.14201) 361 : audit [DBG] from='client.15000 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:11.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:10 vm07 bash[17031]: audit 2026-03-07T07:01:10.306240+0000 mgr.vm07.yrfcuj (mgr.14201) 361 : audit [DBG] from='client.15000 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:11.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:10 vm07 bash[17031]: audit 2026-03-07T07:01:10.473759+0000 mgr.vm07.yrfcuj (mgr.14201) 362 : audit [DBG] from='client.24641 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:11.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:10 vm07 bash[17031]: audit 2026-03-07T07:01:10.473759+0000 mgr.vm07.yrfcuj (mgr.14201) 362 : audit [DBG] from='client.24641 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:11.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:10 vm07 bash[17031]: audit 2026-03-07T07:01:10.657587+0000 mon.vm07 (mon.0) 881 : audit [DBG] from='client.? 192.168.123.107:0/1485722833' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:11.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:10 vm07 bash[17031]: audit 2026-03-07T07:01:10.657587+0000 mon.vm07 (mon.0) 881 : audit [DBG] from='client.? 192.168.123.107:0/1485722833' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:12.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:11 vm09 bash[21220]: cluster 2026-03-07T07:01:10.595894+0000 mgr.vm07.yrfcuj (mgr.14201) 363 : cluster [DBG] pgmap v225: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:12.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:11 vm09 bash[21220]: cluster 2026-03-07T07:01:10.595894+0000 mgr.vm07.yrfcuj (mgr.14201) 363 : cluster [DBG] pgmap v225: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:12.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:11 vm07 bash[17031]: cluster 2026-03-07T07:01:10.595894+0000 mgr.vm07.yrfcuj (mgr.14201) 363 : cluster [DBG] pgmap v225: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:12.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:11 vm07 bash[17031]: cluster 2026-03-07T07:01:10.595894+0000 mgr.vm07.yrfcuj (mgr.14201) 363 : cluster [DBG] pgmap v225: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:14.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:13 vm09 bash[21220]: cluster 2026-03-07T07:01:12.596302+0000 mgr.vm07.yrfcuj (mgr.14201) 364 : cluster [DBG] pgmap v226: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:14.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:13 vm09 bash[21220]: cluster 2026-03-07T07:01:12.596302+0000 mgr.vm07.yrfcuj (mgr.14201) 364 : cluster [DBG] pgmap v226: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:14.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:13 vm07 bash[17031]: cluster 2026-03-07T07:01:12.596302+0000 mgr.vm07.yrfcuj (mgr.14201) 364 : cluster [DBG] pgmap v226: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:14.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:13 vm07 bash[17031]: cluster 2026-03-07T07:01:12.596302+0000 mgr.vm07.yrfcuj (mgr.14201) 364 : cluster [DBG] pgmap v226: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:15.834 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:01:15.985 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:01:15.985 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 19s ago 3m - - 2026-03-07T08:01:15.985 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 19s ago 3m 105M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:01:15.985 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 19s ago 3m 98.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:01:15.985 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (3m) 19s ago 3m 98.9M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:01:16.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:15 vm09 bash[21220]: cluster 2026-03-07T07:01:14.596683+0000 mgr.vm07.yrfcuj (mgr.14201) 365 : cluster [DBG] pgmap v227: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:16.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:15 vm09 bash[21220]: cluster 2026-03-07T07:01:14.596683+0000 mgr.vm07.yrfcuj (mgr.14201) 365 : cluster [DBG] pgmap v227: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:16.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:15 vm07 bash[17031]: cluster 2026-03-07T07:01:14.596683+0000 mgr.vm07.yrfcuj (mgr.14201) 365 : cluster [DBG] pgmap v227: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:16.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:15 vm07 bash[17031]: cluster 2026-03-07T07:01:14.596683+0000 mgr.vm07.yrfcuj (mgr.14201) 365 : cluster [DBG] pgmap v227: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:16.183 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:01:16.183 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:01:16.183 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:01:17.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:16 vm09 bash[21220]: audit 2026-03-07T07:01:15.822092+0000 mgr.vm07.yrfcuj (mgr.14201) 366 : audit [DBG] from='client.15012 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:17.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:16 vm09 bash[21220]: audit 2026-03-07T07:01:15.822092+0000 mgr.vm07.yrfcuj (mgr.14201) 366 : audit [DBG] from='client.15012 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:17.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:16 vm09 bash[21220]: audit 2026-03-07T07:01:15.987218+0000 mgr.vm07.yrfcuj (mgr.14201) 367 : audit [DBG] from='client.24649 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:17.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:16 vm09 bash[21220]: audit 2026-03-07T07:01:15.987218+0000 mgr.vm07.yrfcuj (mgr.14201) 367 : audit [DBG] from='client.24649 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:17.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:16 vm09 bash[21220]: audit 2026-03-07T07:01:16.187248+0000 mon.vm07 (mon.0) 882 : audit [DBG] from='client.? 192.168.123.107:0/2607284409' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:17.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:16 vm09 bash[21220]: audit 2026-03-07T07:01:16.187248+0000 mon.vm07 (mon.0) 882 : audit [DBG] from='client.? 192.168.123.107:0/2607284409' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:17.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:16 vm07 bash[17031]: audit 2026-03-07T07:01:15.822092+0000 mgr.vm07.yrfcuj (mgr.14201) 366 : audit [DBG] from='client.15012 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:17.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:16 vm07 bash[17031]: audit 2026-03-07T07:01:15.822092+0000 mgr.vm07.yrfcuj (mgr.14201) 366 : audit [DBG] from='client.15012 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:17.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:16 vm07 bash[17031]: audit 2026-03-07T07:01:15.987218+0000 mgr.vm07.yrfcuj (mgr.14201) 367 : audit [DBG] from='client.24649 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:17.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:16 vm07 bash[17031]: audit 2026-03-07T07:01:15.987218+0000 mgr.vm07.yrfcuj (mgr.14201) 367 : audit [DBG] from='client.24649 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:17.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:16 vm07 bash[17031]: audit 2026-03-07T07:01:16.187248+0000 mon.vm07 (mon.0) 882 : audit [DBG] from='client.? 192.168.123.107:0/2607284409' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:17.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:16 vm07 bash[17031]: audit 2026-03-07T07:01:16.187248+0000 mon.vm07 (mon.0) 882 : audit [DBG] from='client.? 192.168.123.107:0/2607284409' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:18.081 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:17 vm09 bash[21220]: cluster 2026-03-07T07:01:16.597055+0000 mgr.vm07.yrfcuj (mgr.14201) 368 : cluster [DBG] pgmap v228: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:18.081 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:17 vm09 bash[21220]: cluster 2026-03-07T07:01:16.597055+0000 mgr.vm07.yrfcuj (mgr.14201) 368 : cluster [DBG] pgmap v228: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:18.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:17 vm07 bash[17031]: cluster 2026-03-07T07:01:16.597055+0000 mgr.vm07.yrfcuj (mgr.14201) 368 : cluster [DBG] pgmap v228: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:18.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:17 vm07 bash[17031]: cluster 2026-03-07T07:01:16.597055+0000 mgr.vm07.yrfcuj (mgr.14201) 368 : cluster [DBG] pgmap v228: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:20.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:19 vm09 bash[21220]: cluster 2026-03-07T07:01:18.597516+0000 mgr.vm07.yrfcuj (mgr.14201) 369 : cluster [DBG] pgmap v229: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:20.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:19 vm09 bash[21220]: cluster 2026-03-07T07:01:18.597516+0000 mgr.vm07.yrfcuj (mgr.14201) 369 : cluster [DBG] pgmap v229: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:20.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:19 vm07 bash[17031]: cluster 2026-03-07T07:01:18.597516+0000 mgr.vm07.yrfcuj (mgr.14201) 369 : cluster [DBG] pgmap v229: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:20.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:19 vm07 bash[17031]: cluster 2026-03-07T07:01:18.597516+0000 mgr.vm07.yrfcuj (mgr.14201) 369 : cluster [DBG] pgmap v229: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:21.360 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:01:21.515 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:01:21.515 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 25s ago 3m - - 2026-03-07T08:01:21.515 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 25s ago 3m 105M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:01:21.515 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 25s ago 3m 98.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:01:21.515 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (3m) 25s ago 3m 98.9M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:01:21.709 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:01:21.709 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:01:21.709 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:01:22.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:21 vm09 bash[21220]: cluster 2026-03-07T07:01:20.597972+0000 mgr.vm07.yrfcuj (mgr.14201) 370 : cluster [DBG] pgmap v230: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:22.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:21 vm09 bash[21220]: cluster 2026-03-07T07:01:20.597972+0000 mgr.vm07.yrfcuj (mgr.14201) 370 : cluster [DBG] pgmap v230: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:22.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:21 vm09 bash[21220]: audit 2026-03-07T07:01:21.349173+0000 mgr.vm07.yrfcuj (mgr.14201) 371 : audit [DBG] from='client.15024 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:22.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:21 vm09 bash[21220]: audit 2026-03-07T07:01:21.349173+0000 mgr.vm07.yrfcuj (mgr.14201) 371 : audit [DBG] from='client.15024 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:22.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:21 vm09 bash[21220]: audit 2026-03-07T07:01:21.517333+0000 mgr.vm07.yrfcuj (mgr.14201) 372 : audit [DBG] from='client.15028 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:22.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:21 vm09 bash[21220]: audit 2026-03-07T07:01:21.517333+0000 mgr.vm07.yrfcuj (mgr.14201) 372 : audit [DBG] from='client.15028 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:22.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:21 vm09 bash[21220]: audit 2026-03-07T07:01:21.713207+0000 mon.vm07 (mon.0) 883 : audit [DBG] from='client.? 192.168.123.107:0/3871296208' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:22.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:21 vm09 bash[21220]: audit 2026-03-07T07:01:21.713207+0000 mon.vm07 (mon.0) 883 : audit [DBG] from='client.? 192.168.123.107:0/3871296208' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:22.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:21 vm07 bash[17031]: cluster 2026-03-07T07:01:20.597972+0000 mgr.vm07.yrfcuj (mgr.14201) 370 : cluster [DBG] pgmap v230: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:22.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:21 vm07 bash[17031]: cluster 2026-03-07T07:01:20.597972+0000 mgr.vm07.yrfcuj (mgr.14201) 370 : cluster [DBG] pgmap v230: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:22.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:21 vm07 bash[17031]: audit 2026-03-07T07:01:21.349173+0000 mgr.vm07.yrfcuj (mgr.14201) 371 : audit [DBG] from='client.15024 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:22.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:21 vm07 bash[17031]: audit 2026-03-07T07:01:21.349173+0000 mgr.vm07.yrfcuj (mgr.14201) 371 : audit [DBG] from='client.15024 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:22.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:21 vm07 bash[17031]: audit 2026-03-07T07:01:21.517333+0000 mgr.vm07.yrfcuj (mgr.14201) 372 : audit [DBG] from='client.15028 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:22.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:21 vm07 bash[17031]: audit 2026-03-07T07:01:21.517333+0000 mgr.vm07.yrfcuj (mgr.14201) 372 : audit [DBG] from='client.15028 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:22.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:21 vm07 bash[17031]: audit 2026-03-07T07:01:21.713207+0000 mon.vm07 (mon.0) 883 : audit [DBG] from='client.? 192.168.123.107:0/3871296208' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:22.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:21 vm07 bash[17031]: audit 2026-03-07T07:01:21.713207+0000 mon.vm07 (mon.0) 883 : audit [DBG] from='client.? 192.168.123.107:0/3871296208' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:24.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:23 vm09 bash[21220]: cluster 2026-03-07T07:01:22.598400+0000 mgr.vm07.yrfcuj (mgr.14201) 373 : cluster [DBG] pgmap v231: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:24.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:23 vm09 bash[21220]: cluster 2026-03-07T07:01:22.598400+0000 mgr.vm07.yrfcuj (mgr.14201) 373 : cluster [DBG] pgmap v231: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:24.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:23 vm07 bash[17031]: cluster 2026-03-07T07:01:22.598400+0000 mgr.vm07.yrfcuj (mgr.14201) 373 : cluster [DBG] pgmap v231: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:24.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:23 vm07 bash[17031]: cluster 2026-03-07T07:01:22.598400+0000 mgr.vm07.yrfcuj (mgr.14201) 373 : cluster [DBG] pgmap v231: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:26.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:25 vm09 bash[21220]: cluster 2026-03-07T07:01:24.598815+0000 mgr.vm07.yrfcuj (mgr.14201) 374 : cluster [DBG] pgmap v232: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:26.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:25 vm09 bash[21220]: cluster 2026-03-07T07:01:24.598815+0000 mgr.vm07.yrfcuj (mgr.14201) 374 : cluster [DBG] pgmap v232: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:26.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:25 vm09 bash[21220]: audit 2026-03-07T07:01:25.131131+0000 mon.vm07 (mon.0) 884 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:01:26.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:25 vm09 bash[21220]: audit 2026-03-07T07:01:25.131131+0000 mon.vm07 (mon.0) 884 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:01:26.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:25 vm07 bash[17031]: cluster 2026-03-07T07:01:24.598815+0000 mgr.vm07.yrfcuj (mgr.14201) 374 : cluster [DBG] pgmap v232: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:26.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:25 vm07 bash[17031]: cluster 2026-03-07T07:01:24.598815+0000 mgr.vm07.yrfcuj (mgr.14201) 374 : cluster [DBG] pgmap v232: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:26.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:25 vm07 bash[17031]: audit 2026-03-07T07:01:25.131131+0000 mon.vm07 (mon.0) 884 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:01:26.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:25 vm07 bash[17031]: audit 2026-03-07T07:01:25.131131+0000 mon.vm07 (mon.0) 884 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:01:26.884 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:01:27.037 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:01:27.038 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 30s ago 3m - - 2026-03-07T08:01:27.038 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 30s ago 3m 105M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:01:27.038 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 31s ago 3m 98.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:01:27.038 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (3m) 31s ago 3m 98.9M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:01:27.237 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:01:27.237 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:01:27.237 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:01:28.081 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:27 vm09 bash[21220]: cluster 2026-03-07T07:01:26.599169+0000 mgr.vm07.yrfcuj (mgr.14201) 375 : cluster [DBG] pgmap v233: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:28.081 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:27 vm09 bash[21220]: cluster 2026-03-07T07:01:26.599169+0000 mgr.vm07.yrfcuj (mgr.14201) 375 : cluster [DBG] pgmap v233: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:28.081 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:27 vm09 bash[21220]: audit 2026-03-07T07:01:26.873913+0000 mgr.vm07.yrfcuj (mgr.14201) 376 : audit [DBG] from='client.15036 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:28.081 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:27 vm09 bash[21220]: audit 2026-03-07T07:01:26.873913+0000 mgr.vm07.yrfcuj (mgr.14201) 376 : audit [DBG] from='client.15036 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:28.081 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:27 vm09 bash[21220]: audit 2026-03-07T07:01:27.039225+0000 mgr.vm07.yrfcuj (mgr.14201) 377 : audit [DBG] from='client.15040 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:28.081 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:27 vm09 bash[21220]: audit 2026-03-07T07:01:27.039225+0000 mgr.vm07.yrfcuj (mgr.14201) 377 : audit [DBG] from='client.15040 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:28.081 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:27 vm09 bash[21220]: audit 2026-03-07T07:01:27.241057+0000 mon.vm07 (mon.0) 885 : audit [DBG] from='client.? 192.168.123.107:0/3919655700' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:28.081 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:27 vm09 bash[21220]: audit 2026-03-07T07:01:27.241057+0000 mon.vm07 (mon.0) 885 : audit [DBG] from='client.? 192.168.123.107:0/3919655700' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:28.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:27 vm07 bash[17031]: cluster 2026-03-07T07:01:26.599169+0000 mgr.vm07.yrfcuj (mgr.14201) 375 : cluster [DBG] pgmap v233: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:28.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:27 vm07 bash[17031]: cluster 2026-03-07T07:01:26.599169+0000 mgr.vm07.yrfcuj (mgr.14201) 375 : cluster [DBG] pgmap v233: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:28.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:27 vm07 bash[17031]: audit 2026-03-07T07:01:26.873913+0000 mgr.vm07.yrfcuj (mgr.14201) 376 : audit [DBG] from='client.15036 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:28.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:27 vm07 bash[17031]: audit 2026-03-07T07:01:26.873913+0000 mgr.vm07.yrfcuj (mgr.14201) 376 : audit [DBG] from='client.15036 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:28.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:27 vm07 bash[17031]: audit 2026-03-07T07:01:27.039225+0000 mgr.vm07.yrfcuj (mgr.14201) 377 : audit [DBG] from='client.15040 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:28.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:27 vm07 bash[17031]: audit 2026-03-07T07:01:27.039225+0000 mgr.vm07.yrfcuj (mgr.14201) 377 : audit [DBG] from='client.15040 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:28.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:27 vm07 bash[17031]: audit 2026-03-07T07:01:27.241057+0000 mon.vm07 (mon.0) 885 : audit [DBG] from='client.? 192.168.123.107:0/3919655700' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:28.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:27 vm07 bash[17031]: audit 2026-03-07T07:01:27.241057+0000 mon.vm07 (mon.0) 885 : audit [DBG] from='client.? 192.168.123.107:0/3919655700' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:30.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:29 vm09 bash[21220]: cluster 2026-03-07T07:01:28.599595+0000 mgr.vm07.yrfcuj (mgr.14201) 378 : cluster [DBG] pgmap v234: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:29 vm09 bash[21220]: cluster 2026-03-07T07:01:28.599595+0000 mgr.vm07.yrfcuj (mgr.14201) 378 : cluster [DBG] pgmap v234: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:30.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:29 vm07 bash[17031]: cluster 2026-03-07T07:01:28.599595+0000 mgr.vm07.yrfcuj (mgr.14201) 378 : cluster [DBG] pgmap v234: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:30.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:29 vm07 bash[17031]: cluster 2026-03-07T07:01:28.599595+0000 mgr.vm07.yrfcuj (mgr.14201) 378 : cluster [DBG] pgmap v234: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:32.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:31 vm09 bash[21220]: cluster 2026-03-07T07:01:30.600023+0000 mgr.vm07.yrfcuj (mgr.14201) 379 : cluster [DBG] pgmap v235: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:32.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:31 vm09 bash[21220]: cluster 2026-03-07T07:01:30.600023+0000 mgr.vm07.yrfcuj (mgr.14201) 379 : cluster [DBG] pgmap v235: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:32.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:31 vm07 bash[17031]: cluster 2026-03-07T07:01:30.600023+0000 mgr.vm07.yrfcuj (mgr.14201) 379 : cluster [DBG] pgmap v235: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:32.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:31 vm07 bash[17031]: cluster 2026-03-07T07:01:30.600023+0000 mgr.vm07.yrfcuj (mgr.14201) 379 : cluster [DBG] pgmap v235: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:32.424 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:01:32.582 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:01:32.582 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 36s ago 3m - - 2026-03-07T08:01:32.582 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 36s ago 3m 105M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:01:32.582 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 36s ago 3m 98.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:01:32.582 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (3m) 36s ago 3m 98.9M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:01:32.781 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:01:32.781 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:01:32.781 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:01:33.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:32 vm09 bash[21220]: audit 2026-03-07T07:01:32.411311+0000 mgr.vm07.yrfcuj (mgr.14201) 380 : audit [DBG] from='client.15048 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:33.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:32 vm09 bash[21220]: audit 2026-03-07T07:01:32.411311+0000 mgr.vm07.yrfcuj (mgr.14201) 380 : audit [DBG] from='client.15048 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:33.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:32 vm09 bash[21220]: audit 2026-03-07T07:01:32.784616+0000 mon.vm07 (mon.0) 886 : audit [DBG] from='client.? 192.168.123.107:0/3838550317' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:33.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:32 vm09 bash[21220]: audit 2026-03-07T07:01:32.784616+0000 mon.vm07 (mon.0) 886 : audit [DBG] from='client.? 192.168.123.107:0/3838550317' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:33.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:32 vm07 bash[17031]: audit 2026-03-07T07:01:32.411311+0000 mgr.vm07.yrfcuj (mgr.14201) 380 : audit [DBG] from='client.15048 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:33.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:32 vm07 bash[17031]: audit 2026-03-07T07:01:32.411311+0000 mgr.vm07.yrfcuj (mgr.14201) 380 : audit [DBG] from='client.15048 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:33.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:32 vm07 bash[17031]: audit 2026-03-07T07:01:32.784616+0000 mon.vm07 (mon.0) 886 : audit [DBG] from='client.? 192.168.123.107:0/3838550317' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:33.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:32 vm07 bash[17031]: audit 2026-03-07T07:01:32.784616+0000 mon.vm07 (mon.0) 886 : audit [DBG] from='client.? 192.168.123.107:0/3838550317' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:33 vm09 bash[21220]: audit 2026-03-07T07:01:32.582546+0000 mgr.vm07.yrfcuj (mgr.14201) 381 : audit [DBG] from='client.15052 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:33 vm09 bash[21220]: audit 2026-03-07T07:01:32.582546+0000 mgr.vm07.yrfcuj (mgr.14201) 381 : audit [DBG] from='client.15052 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:33 vm09 bash[21220]: cluster 2026-03-07T07:01:32.600404+0000 mgr.vm07.yrfcuj (mgr.14201) 382 : cluster [DBG] pgmap v236: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:33 vm09 bash[21220]: cluster 2026-03-07T07:01:32.600404+0000 mgr.vm07.yrfcuj (mgr.14201) 382 : cluster [DBG] pgmap v236: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:34.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:33 vm07 bash[17031]: audit 2026-03-07T07:01:32.582546+0000 mgr.vm07.yrfcuj (mgr.14201) 381 : audit [DBG] from='client.15052 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:34.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:33 vm07 bash[17031]: audit 2026-03-07T07:01:32.582546+0000 mgr.vm07.yrfcuj (mgr.14201) 381 : audit [DBG] from='client.15052 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:34.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:33 vm07 bash[17031]: cluster 2026-03-07T07:01:32.600404+0000 mgr.vm07.yrfcuj (mgr.14201) 382 : cluster [DBG] pgmap v236: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:34.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:33 vm07 bash[17031]: cluster 2026-03-07T07:01:32.600404+0000 mgr.vm07.yrfcuj (mgr.14201) 382 : cluster [DBG] pgmap v236: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:35 vm09 bash[21220]: cluster 2026-03-07T07:01:34.600879+0000 mgr.vm07.yrfcuj (mgr.14201) 383 : cluster [DBG] pgmap v237: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:35 vm09 bash[21220]: cluster 2026-03-07T07:01:34.600879+0000 mgr.vm07.yrfcuj (mgr.14201) 383 : cluster [DBG] pgmap v237: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:36.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:35 vm07 bash[17031]: cluster 2026-03-07T07:01:34.600879+0000 mgr.vm07.yrfcuj (mgr.14201) 383 : cluster [DBG] pgmap v237: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:36.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:35 vm07 bash[17031]: cluster 2026-03-07T07:01:34.600879+0000 mgr.vm07.yrfcuj (mgr.14201) 383 : cluster [DBG] pgmap v237: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:37.953 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:01:38.107 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:01:38.107 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 41s ago 4m - - 2026-03-07T08:01:38.107 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 41s ago 3m 105M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:01:38.108 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 42s ago 4m 98.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:01:38.108 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (3m) 42s ago 3m 98.9M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:01:38.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:37 vm07 bash[17031]: cluster 2026-03-07T07:01:36.601256+0000 mgr.vm07.yrfcuj (mgr.14201) 384 : cluster [DBG] pgmap v238: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:38.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:37 vm07 bash[17031]: cluster 2026-03-07T07:01:36.601256+0000 mgr.vm07.yrfcuj (mgr.14201) 384 : cluster [DBG] pgmap v238: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:38.296 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:01:38.296 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:01:38.296 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:01:38.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:37 vm09 bash[21220]: cluster 2026-03-07T07:01:36.601256+0000 mgr.vm07.yrfcuj (mgr.14201) 384 : cluster [DBG] pgmap v238: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:38.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:37 vm09 bash[21220]: cluster 2026-03-07T07:01:36.601256+0000 mgr.vm07.yrfcuj (mgr.14201) 384 : cluster [DBG] pgmap v238: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:39.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:38 vm07 bash[17031]: audit 2026-03-07T07:01:37.941502+0000 mgr.vm07.yrfcuj (mgr.14201) 385 : audit [DBG] from='client.15060 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:39.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:38 vm07 bash[17031]: audit 2026-03-07T07:01:37.941502+0000 mgr.vm07.yrfcuj (mgr.14201) 385 : audit [DBG] from='client.15060 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:39.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:38 vm07 bash[17031]: audit 2026-03-07T07:01:38.107487+0000 mgr.vm07.yrfcuj (mgr.14201) 386 : audit [DBG] from='client.15064 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:39.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:38 vm07 bash[17031]: audit 2026-03-07T07:01:38.107487+0000 mgr.vm07.yrfcuj (mgr.14201) 386 : audit [DBG] from='client.15064 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:39.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:38 vm07 bash[17031]: audit 2026-03-07T07:01:38.298570+0000 mon.vm07 (mon.0) 887 : audit [DBG] from='client.? 192.168.123.107:0/1732238201' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:39.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:38 vm07 bash[17031]: audit 2026-03-07T07:01:38.298570+0000 mon.vm07 (mon.0) 887 : audit [DBG] from='client.? 192.168.123.107:0/1732238201' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:39.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:38 vm09 bash[21220]: audit 2026-03-07T07:01:37.941502+0000 mgr.vm07.yrfcuj (mgr.14201) 385 : audit [DBG] from='client.15060 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:39.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:38 vm09 bash[21220]: audit 2026-03-07T07:01:37.941502+0000 mgr.vm07.yrfcuj (mgr.14201) 385 : audit [DBG] from='client.15060 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:39.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:38 vm09 bash[21220]: audit 2026-03-07T07:01:38.107487+0000 mgr.vm07.yrfcuj (mgr.14201) 386 : audit [DBG] from='client.15064 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:39.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:38 vm09 bash[21220]: audit 2026-03-07T07:01:38.107487+0000 mgr.vm07.yrfcuj (mgr.14201) 386 : audit [DBG] from='client.15064 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:39.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:38 vm09 bash[21220]: audit 2026-03-07T07:01:38.298570+0000 mon.vm07 (mon.0) 887 : audit [DBG] from='client.? 192.168.123.107:0/1732238201' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:39.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:38 vm09 bash[21220]: audit 2026-03-07T07:01:38.298570+0000 mon.vm07 (mon.0) 887 : audit [DBG] from='client.? 192.168.123.107:0/1732238201' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:40.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:39 vm07 bash[17031]: cluster 2026-03-07T07:01:38.601815+0000 mgr.vm07.yrfcuj (mgr.14201) 387 : cluster [DBG] pgmap v239: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:40.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:39 vm07 bash[17031]: cluster 2026-03-07T07:01:38.601815+0000 mgr.vm07.yrfcuj (mgr.14201) 387 : cluster [DBG] pgmap v239: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:40.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:39 vm09 bash[21220]: cluster 2026-03-07T07:01:38.601815+0000 mgr.vm07.yrfcuj (mgr.14201) 387 : cluster [DBG] pgmap v239: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:40.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:39 vm09 bash[21220]: cluster 2026-03-07T07:01:38.601815+0000 mgr.vm07.yrfcuj (mgr.14201) 387 : cluster [DBG] pgmap v239: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:41.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:40 vm07 bash[17031]: audit 2026-03-07T07:01:40.131180+0000 mon.vm07 (mon.0) 888 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:01:41.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:40 vm07 bash[17031]: audit 2026-03-07T07:01:40.131180+0000 mon.vm07 (mon.0) 888 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:01:41.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:40 vm09 bash[21220]: audit 2026-03-07T07:01:40.131180+0000 mon.vm07 (mon.0) 888 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:01:41.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:40 vm09 bash[21220]: audit 2026-03-07T07:01:40.131180+0000 mon.vm07 (mon.0) 888 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:01:42.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:41 vm07 bash[17031]: cluster 2026-03-07T07:01:40.602161+0000 mgr.vm07.yrfcuj (mgr.14201) 388 : cluster [DBG] pgmap v240: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:42.151 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:41 vm07 bash[17031]: cluster 2026-03-07T07:01:40.602161+0000 mgr.vm07.yrfcuj (mgr.14201) 388 : cluster [DBG] pgmap v240: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:42.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:41 vm09 bash[21220]: cluster 2026-03-07T07:01:40.602161+0000 mgr.vm07.yrfcuj (mgr.14201) 388 : cluster [DBG] pgmap v240: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:42.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:41 vm09 bash[21220]: cluster 2026-03-07T07:01:40.602161+0000 mgr.vm07.yrfcuj (mgr.14201) 388 : cluster [DBG] pgmap v240: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:43.466 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:01:43.615 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:01:43.615 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 47s ago 4m - - 2026-03-07T08:01:43.615 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 47s ago 4m 105M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:01:43.615 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 47s ago 4m 98.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:01:43.615 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (4m) 47s ago 4m 98.9M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:01:43.816 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:01:43.816 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:01:43.816 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:01:44.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:43 vm09 bash[21220]: cluster 2026-03-07T07:01:42.602515+0000 mgr.vm07.yrfcuj (mgr.14201) 389 : cluster [DBG] pgmap v241: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:44.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:43 vm09 bash[21220]: cluster 2026-03-07T07:01:42.602515+0000 mgr.vm07.yrfcuj (mgr.14201) 389 : cluster [DBG] pgmap v241: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:44.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:43 vm09 bash[21220]: audit 2026-03-07T07:01:43.453003+0000 mgr.vm07.yrfcuj (mgr.14201) 390 : audit [DBG] from='client.15072 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:44.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:43 vm09 bash[21220]: audit 2026-03-07T07:01:43.453003+0000 mgr.vm07.yrfcuj (mgr.14201) 390 : audit [DBG] from='client.15072 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:44.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:43 vm09 bash[21220]: audit 2026-03-07T07:01:43.818231+0000 mon.vm07 (mon.0) 889 : audit [DBG] from='client.? 192.168.123.107:0/700319748' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:44.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:43 vm09 bash[21220]: audit 2026-03-07T07:01:43.818231+0000 mon.vm07 (mon.0) 889 : audit [DBG] from='client.? 192.168.123.107:0/700319748' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:44.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:43 vm07 bash[17031]: cluster 2026-03-07T07:01:42.602515+0000 mgr.vm07.yrfcuj (mgr.14201) 389 : cluster [DBG] pgmap v241: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:44.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:43 vm07 bash[17031]: cluster 2026-03-07T07:01:42.602515+0000 mgr.vm07.yrfcuj (mgr.14201) 389 : cluster [DBG] pgmap v241: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:44.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:43 vm07 bash[17031]: audit 2026-03-07T07:01:43.453003+0000 mgr.vm07.yrfcuj (mgr.14201) 390 : audit [DBG] from='client.15072 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:44.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:43 vm07 bash[17031]: audit 2026-03-07T07:01:43.453003+0000 mgr.vm07.yrfcuj (mgr.14201) 390 : audit [DBG] from='client.15072 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:44.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:43 vm07 bash[17031]: audit 2026-03-07T07:01:43.818231+0000 mon.vm07 (mon.0) 889 : audit [DBG] from='client.? 192.168.123.107:0/700319748' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:44.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:43 vm07 bash[17031]: audit 2026-03-07T07:01:43.818231+0000 mon.vm07 (mon.0) 889 : audit [DBG] from='client.? 192.168.123.107:0/700319748' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:44 vm09 bash[21220]: audit 2026-03-07T07:01:43.614215+0000 mgr.vm07.yrfcuj (mgr.14201) 391 : audit [DBG] from='client.15076 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:45.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:44 vm09 bash[21220]: audit 2026-03-07T07:01:43.614215+0000 mgr.vm07.yrfcuj (mgr.14201) 391 : audit [DBG] from='client.15076 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:45.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:44 vm07 bash[17031]: audit 2026-03-07T07:01:43.614215+0000 mgr.vm07.yrfcuj (mgr.14201) 391 : audit [DBG] from='client.15076 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:45.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:44 vm07 bash[17031]: audit 2026-03-07T07:01:43.614215+0000 mgr.vm07.yrfcuj (mgr.14201) 391 : audit [DBG] from='client.15076 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:46.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:45 vm09 bash[21220]: cluster 2026-03-07T07:01:44.602929+0000 mgr.vm07.yrfcuj (mgr.14201) 392 : cluster [DBG] pgmap v242: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:46.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:45 vm09 bash[21220]: cluster 2026-03-07T07:01:44.602929+0000 mgr.vm07.yrfcuj (mgr.14201) 392 : cluster [DBG] pgmap v242: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:46.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:45 vm07 bash[17031]: cluster 2026-03-07T07:01:44.602929+0000 mgr.vm07.yrfcuj (mgr.14201) 392 : cluster [DBG] pgmap v242: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:46.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:45 vm07 bash[17031]: cluster 2026-03-07T07:01:44.602929+0000 mgr.vm07.yrfcuj (mgr.14201) 392 : cluster [DBG] pgmap v242: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:47 vm09 bash[21220]: cluster 2026-03-07T07:01:46.603382+0000 mgr.vm07.yrfcuj (mgr.14201) 393 : cluster [DBG] pgmap v243: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:47 vm09 bash[21220]: cluster 2026-03-07T07:01:46.603382+0000 mgr.vm07.yrfcuj (mgr.14201) 393 : cluster [DBG] pgmap v243: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:48.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:47 vm07 bash[17031]: cluster 2026-03-07T07:01:46.603382+0000 mgr.vm07.yrfcuj (mgr.14201) 393 : cluster [DBG] pgmap v243: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:48.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:47 vm07 bash[17031]: cluster 2026-03-07T07:01:46.603382+0000 mgr.vm07.yrfcuj (mgr.14201) 393 : cluster [DBG] pgmap v243: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:01:48.995 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:01:49.152 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:01:49.152 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 52s ago 4m - - 2026-03-07T08:01:49.152 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 52s ago 4m 105M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:01:49.152 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 53s ago 4m 98.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:01:49.152 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (4m) 53s ago 4m 98.9M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:01:49.352 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:01:49.352 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:01:49.352 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:01:50.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:49 vm09 bash[21220]: cluster 2026-03-07T07:01:48.603814+0000 mgr.vm07.yrfcuj (mgr.14201) 394 : cluster [DBG] pgmap v244: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:50.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:49 vm09 bash[21220]: cluster 2026-03-07T07:01:48.603814+0000 mgr.vm07.yrfcuj (mgr.14201) 394 : cluster [DBG] pgmap v244: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:50.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:49 vm09 bash[21220]: audit 2026-03-07T07:01:48.978846+0000 mgr.vm07.yrfcuj (mgr.14201) 395 : audit [DBG] from='client.15084 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:50.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:49 vm09 bash[21220]: audit 2026-03-07T07:01:48.978846+0000 mgr.vm07.yrfcuj (mgr.14201) 395 : audit [DBG] from='client.15084 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:50.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:49 vm09 bash[21220]: audit 2026-03-07T07:01:49.151767+0000 mgr.vm07.yrfcuj (mgr.14201) 396 : audit [DBG] from='client.15088 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:50.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:49 vm09 bash[21220]: audit 2026-03-07T07:01:49.151767+0000 mgr.vm07.yrfcuj (mgr.14201) 396 : audit [DBG] from='client.15088 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:50.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:49 vm09 bash[21220]: audit 2026-03-07T07:01:49.353883+0000 mon.vm07 (mon.0) 890 : audit [DBG] from='client.? 192.168.123.107:0/2462843000' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:50.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:49 vm09 bash[21220]: audit 2026-03-07T07:01:49.353883+0000 mon.vm07 (mon.0) 890 : audit [DBG] from='client.? 192.168.123.107:0/2462843000' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:50.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:49 vm07 bash[17031]: cluster 2026-03-07T07:01:48.603814+0000 mgr.vm07.yrfcuj (mgr.14201) 394 : cluster [DBG] pgmap v244: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:50.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:49 vm07 bash[17031]: cluster 2026-03-07T07:01:48.603814+0000 mgr.vm07.yrfcuj (mgr.14201) 394 : cluster [DBG] pgmap v244: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:50.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:49 vm07 bash[17031]: audit 2026-03-07T07:01:48.978846+0000 mgr.vm07.yrfcuj (mgr.14201) 395 : audit [DBG] from='client.15084 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:50.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:49 vm07 bash[17031]: audit 2026-03-07T07:01:48.978846+0000 mgr.vm07.yrfcuj (mgr.14201) 395 : audit [DBG] from='client.15084 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:50.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:49 vm07 bash[17031]: audit 2026-03-07T07:01:49.151767+0000 mgr.vm07.yrfcuj (mgr.14201) 396 : audit [DBG] from='client.15088 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:50.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:49 vm07 bash[17031]: audit 2026-03-07T07:01:49.151767+0000 mgr.vm07.yrfcuj (mgr.14201) 396 : audit [DBG] from='client.15088 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:50.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:49 vm07 bash[17031]: audit 2026-03-07T07:01:49.353883+0000 mon.vm07 (mon.0) 890 : audit [DBG] from='client.? 192.168.123.107:0/2462843000' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:50.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:49 vm07 bash[17031]: audit 2026-03-07T07:01:49.353883+0000 mon.vm07 (mon.0) 890 : audit [DBG] from='client.? 192.168.123.107:0/2462843000' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:52.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:51 vm09 bash[21220]: cluster 2026-03-07T07:01:50.604234+0000 mgr.vm07.yrfcuj (mgr.14201) 397 : cluster [DBG] pgmap v245: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:52.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:51 vm09 bash[21220]: cluster 2026-03-07T07:01:50.604234+0000 mgr.vm07.yrfcuj (mgr.14201) 397 : cluster [DBG] pgmap v245: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:52.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:51 vm07 bash[17031]: cluster 2026-03-07T07:01:50.604234+0000 mgr.vm07.yrfcuj (mgr.14201) 397 : cluster [DBG] pgmap v245: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:52.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:51 vm07 bash[17031]: cluster 2026-03-07T07:01:50.604234+0000 mgr.vm07.yrfcuj (mgr.14201) 397 : cluster [DBG] pgmap v245: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:54.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:53 vm09 bash[21220]: cluster 2026-03-07T07:01:52.604687+0000 mgr.vm07.yrfcuj (mgr.14201) 398 : cluster [DBG] pgmap v246: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:54.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:53 vm09 bash[21220]: cluster 2026-03-07T07:01:52.604687+0000 mgr.vm07.yrfcuj (mgr.14201) 398 : cluster [DBG] pgmap v246: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:54.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:53 vm07 bash[17031]: cluster 2026-03-07T07:01:52.604687+0000 mgr.vm07.yrfcuj (mgr.14201) 398 : cluster [DBG] pgmap v246: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:54.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:53 vm07 bash[17031]: cluster 2026-03-07T07:01:52.604687+0000 mgr.vm07.yrfcuj (mgr.14201) 398 : cluster [DBG] pgmap v246: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:54.534 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:01:54.706 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:01:54.706 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 58s ago 4m - - 2026-03-07T08:01:54.706 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 58s ago 4m 105M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:01:54.707 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 58s ago 4m 98.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:01:54.707 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (4m) 58s ago 4m 98.9M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:01:54.901 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:01:54.901 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:01:54.901 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:01:55.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:54 vm09 bash[21220]: audit 2026-03-07T07:01:54.518153+0000 mgr.vm07.yrfcuj (mgr.14201) 399 : audit [DBG] from='client.15096 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:55.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:54 vm09 bash[21220]: audit 2026-03-07T07:01:54.518153+0000 mgr.vm07.yrfcuj (mgr.14201) 399 : audit [DBG] from='client.15096 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:55.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:54 vm09 bash[21220]: audit 2026-03-07T07:01:54.901945+0000 mon.vm07 (mon.0) 891 : audit [DBG] from='client.? 192.168.123.107:0/2814575918' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:55.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:54 vm09 bash[21220]: audit 2026-03-07T07:01:54.901945+0000 mon.vm07 (mon.0) 891 : audit [DBG] from='client.? 192.168.123.107:0/2814575918' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:55.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:54 vm07 bash[17031]: audit 2026-03-07T07:01:54.518153+0000 mgr.vm07.yrfcuj (mgr.14201) 399 : audit [DBG] from='client.15096 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:55.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:54 vm07 bash[17031]: audit 2026-03-07T07:01:54.518153+0000 mgr.vm07.yrfcuj (mgr.14201) 399 : audit [DBG] from='client.15096 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:55.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:54 vm07 bash[17031]: audit 2026-03-07T07:01:54.901945+0000 mon.vm07 (mon.0) 891 : audit [DBG] from='client.? 192.168.123.107:0/2814575918' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:55.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:54 vm07 bash[17031]: audit 2026-03-07T07:01:54.901945+0000 mon.vm07 (mon.0) 891 : audit [DBG] from='client.? 192.168.123.107:0/2814575918' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:01:56.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:55 vm09 bash[21220]: cluster 2026-03-07T07:01:54.605087+0000 mgr.vm07.yrfcuj (mgr.14201) 400 : cluster [DBG] pgmap v247: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:56.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:55 vm09 bash[21220]: cluster 2026-03-07T07:01:54.605087+0000 mgr.vm07.yrfcuj (mgr.14201) 400 : cluster [DBG] pgmap v247: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:56.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:55 vm09 bash[21220]: audit 2026-03-07T07:01:54.705135+0000 mgr.vm07.yrfcuj (mgr.14201) 401 : audit [DBG] from='client.15100 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:56.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:55 vm09 bash[21220]: audit 2026-03-07T07:01:54.705135+0000 mgr.vm07.yrfcuj (mgr.14201) 401 : audit [DBG] from='client.15100 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:56.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:55 vm09 bash[21220]: audit 2026-03-07T07:01:55.131584+0000 mon.vm07 (mon.0) 892 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:01:56.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:55 vm09 bash[21220]: audit 2026-03-07T07:01:55.131584+0000 mon.vm07 (mon.0) 892 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:01:56.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:55 vm07 bash[17031]: cluster 2026-03-07T07:01:54.605087+0000 mgr.vm07.yrfcuj (mgr.14201) 400 : cluster [DBG] pgmap v247: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:56.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:55 vm07 bash[17031]: cluster 2026-03-07T07:01:54.605087+0000 mgr.vm07.yrfcuj (mgr.14201) 400 : cluster [DBG] pgmap v247: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:56.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:55 vm07 bash[17031]: audit 2026-03-07T07:01:54.705135+0000 mgr.vm07.yrfcuj (mgr.14201) 401 : audit [DBG] from='client.15100 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:56.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:55 vm07 bash[17031]: audit 2026-03-07T07:01:54.705135+0000 mgr.vm07.yrfcuj (mgr.14201) 401 : audit [DBG] from='client.15100 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:01:56.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:55 vm07 bash[17031]: audit 2026-03-07T07:01:55.131584+0000 mon.vm07 (mon.0) 892 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:01:56.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:55 vm07 bash[17031]: audit 2026-03-07T07:01:55.131584+0000 mon.vm07 (mon.0) 892 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:01:57.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:57 vm09 bash[21220]: audit 2026-03-07T07:01:56.638443+0000 mon.vm07 (mon.0) 893 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:01:57.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:57 vm09 bash[21220]: audit 2026-03-07T07:01:56.638443+0000 mon.vm07 (mon.0) 893 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:01:57.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:57 vm07 bash[17031]: audit 2026-03-07T07:01:56.638443+0000 mon.vm07 (mon.0) 893 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:01:57.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:57 vm07 bash[17031]: audit 2026-03-07T07:01:56.638443+0000 mon.vm07 (mon.0) 893 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:01:58.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:58 vm09 bash[21220]: cluster 2026-03-07T07:01:56.605479+0000 mgr.vm07.yrfcuj (mgr.14201) 402 : cluster [DBG] pgmap v248: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:58.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:01:58 vm09 bash[21220]: cluster 2026-03-07T07:01:56.605479+0000 mgr.vm07.yrfcuj (mgr.14201) 402 : cluster [DBG] pgmap v248: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:58.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:58 vm07 bash[17031]: cluster 2026-03-07T07:01:56.605479+0000 mgr.vm07.yrfcuj (mgr.14201) 402 : cluster [DBG] pgmap v248: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:01:58.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:01:58 vm07 bash[17031]: cluster 2026-03-07T07:01:56.605479+0000 mgr.vm07.yrfcuj (mgr.14201) 402 : cluster [DBG] pgmap v248: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:00.093 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:02:00.246 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:02:00.246 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 63s ago 4m - - 2026-03-07T08:02:00.246 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 63s ago 4m 105M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:02:00.246 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 64s ago 4m 98.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:02:00.246 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (4m) 64s ago 4m 98.9M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:02:00.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:00 vm09 bash[21220]: cluster 2026-03-07T07:01:58.605998+0000 mgr.vm07.yrfcuj (mgr.14201) 403 : cluster [DBG] pgmap v249: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:00.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:00 vm09 bash[21220]: cluster 2026-03-07T07:01:58.605998+0000 mgr.vm07.yrfcuj (mgr.14201) 403 : cluster [DBG] pgmap v249: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:00.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:00 vm07 bash[17031]: cluster 2026-03-07T07:01:58.605998+0000 mgr.vm07.yrfcuj (mgr.14201) 403 : cluster [DBG] pgmap v249: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:00.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:00 vm07 bash[17031]: cluster 2026-03-07T07:01:58.605998+0000 mgr.vm07.yrfcuj (mgr.14201) 403 : cluster [DBG] pgmap v249: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:00.436 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:02:00.436 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:02:00.436 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:02:01.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:01 vm09 bash[21220]: audit 2026-03-07T07:02:00.080211+0000 mgr.vm07.yrfcuj (mgr.14201) 404 : audit [DBG] from='client.15108 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:01.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:01 vm09 bash[21220]: audit 2026-03-07T07:02:00.080211+0000 mgr.vm07.yrfcuj (mgr.14201) 404 : audit [DBG] from='client.15108 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:01.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:01 vm09 bash[21220]: audit 2026-03-07T07:02:00.245036+0000 mgr.vm07.yrfcuj (mgr.14201) 405 : audit [DBG] from='client.15112 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:01.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:01 vm09 bash[21220]: audit 2026-03-07T07:02:00.245036+0000 mgr.vm07.yrfcuj (mgr.14201) 405 : audit [DBG] from='client.15112 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:01.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:01 vm09 bash[21220]: audit 2026-03-07T07:02:00.436912+0000 mon.vm07 (mon.0) 894 : audit [DBG] from='client.? 192.168.123.107:0/2358646538' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:01.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:01 vm09 bash[21220]: audit 2026-03-07T07:02:00.436912+0000 mon.vm07 (mon.0) 894 : audit [DBG] from='client.? 192.168.123.107:0/2358646538' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:01.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:01 vm07 bash[17031]: audit 2026-03-07T07:02:00.080211+0000 mgr.vm07.yrfcuj (mgr.14201) 404 : audit [DBG] from='client.15108 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:01.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:01 vm07 bash[17031]: audit 2026-03-07T07:02:00.080211+0000 mgr.vm07.yrfcuj (mgr.14201) 404 : audit [DBG] from='client.15108 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:01.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:01 vm07 bash[17031]: audit 2026-03-07T07:02:00.245036+0000 mgr.vm07.yrfcuj (mgr.14201) 405 : audit [DBG] from='client.15112 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:01.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:01 vm07 bash[17031]: audit 2026-03-07T07:02:00.245036+0000 mgr.vm07.yrfcuj (mgr.14201) 405 : audit [DBG] from='client.15112 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:01.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:01 vm07 bash[17031]: audit 2026-03-07T07:02:00.436912+0000 mon.vm07 (mon.0) 894 : audit [DBG] from='client.? 192.168.123.107:0/2358646538' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:01.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:01 vm07 bash[17031]: audit 2026-03-07T07:02:00.436912+0000 mon.vm07 (mon.0) 894 : audit [DBG] from='client.? 192.168.123.107:0/2358646538' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:02.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:02 vm09 bash[21220]: cluster 2026-03-07T07:02:00.606425+0000 mgr.vm07.yrfcuj (mgr.14201) 406 : cluster [DBG] pgmap v250: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:02.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:02 vm09 bash[21220]: cluster 2026-03-07T07:02:00.606425+0000 mgr.vm07.yrfcuj (mgr.14201) 406 : cluster [DBG] pgmap v250: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:02.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:02 vm09 bash[21220]: audit 2026-03-07T07:02:01.939612+0000 mon.vm07 (mon.0) 895 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:02.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:02 vm09 bash[21220]: audit 2026-03-07T07:02:01.939612+0000 mon.vm07 (mon.0) 895 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:02.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:02 vm09 bash[21220]: audit 2026-03-07T07:02:01.944972+0000 mon.vm07 (mon.0) 896 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:02.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:02 vm09 bash[21220]: audit 2026-03-07T07:02:01.944972+0000 mon.vm07 (mon.0) 896 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:02.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:02 vm07 bash[17031]: cluster 2026-03-07T07:02:00.606425+0000 mgr.vm07.yrfcuj (mgr.14201) 406 : cluster [DBG] pgmap v250: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:02.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:02 vm07 bash[17031]: cluster 2026-03-07T07:02:00.606425+0000 mgr.vm07.yrfcuj (mgr.14201) 406 : cluster [DBG] pgmap v250: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:02.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:02 vm07 bash[17031]: audit 2026-03-07T07:02:01.939612+0000 mon.vm07 (mon.0) 895 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:02.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:02 vm07 bash[17031]: audit 2026-03-07T07:02:01.939612+0000 mon.vm07 (mon.0) 895 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:02.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:02 vm07 bash[17031]: audit 2026-03-07T07:02:01.944972+0000 mon.vm07 (mon.0) 896 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:02.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:02 vm07 bash[17031]: audit 2026-03-07T07:02:01.944972+0000 mon.vm07 (mon.0) 896 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:03 vm09 bash[21220]: audit 2026-03-07T07:02:02.184841+0000 mon.vm07 (mon.0) 897 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:03 vm09 bash[21220]: audit 2026-03-07T07:02:02.184841+0000 mon.vm07 (mon.0) 897 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:03 vm09 bash[21220]: audit 2026-03-07T07:02:02.190104+0000 mon.vm07 (mon.0) 898 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:03 vm09 bash[21220]: audit 2026-03-07T07:02:02.190104+0000 mon.vm07 (mon.0) 898 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:03 vm09 bash[21220]: audit 2026-03-07T07:02:02.484574+0000 mon.vm07 (mon.0) 899 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:02:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:03 vm09 bash[21220]: audit 2026-03-07T07:02:02.484574+0000 mon.vm07 (mon.0) 899 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:02:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:03 vm09 bash[21220]: audit 2026-03-07T07:02:02.485136+0000 mon.vm07 (mon.0) 900 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:02:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:03 vm09 bash[21220]: audit 2026-03-07T07:02:02.485136+0000 mon.vm07 (mon.0) 900 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:02:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:03 vm09 bash[21220]: cluster 2026-03-07T07:02:02.486021+0000 mgr.vm07.yrfcuj (mgr.14201) 407 : cluster [DBG] pgmap v251: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:03 vm09 bash[21220]: cluster 2026-03-07T07:02:02.486021+0000 mgr.vm07.yrfcuj (mgr.14201) 407 : cluster [DBG] pgmap v251: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:03 vm09 bash[21220]: cluster 2026-03-07T07:02:02.486185+0000 mgr.vm07.yrfcuj (mgr.14201) 408 : cluster [DBG] pgmap v252: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:03 vm09 bash[21220]: cluster 2026-03-07T07:02:02.486185+0000 mgr.vm07.yrfcuj (mgr.14201) 408 : cluster [DBG] pgmap v252: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:03 vm09 bash[21220]: audit 2026-03-07T07:02:02.489897+0000 mon.vm07 (mon.0) 901 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:03 vm09 bash[21220]: audit 2026-03-07T07:02:02.489897+0000 mon.vm07 (mon.0) 901 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:03 vm09 bash[21220]: audit 2026-03-07T07:02:02.491183+0000 mon.vm07 (mon.0) 902 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:02:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:03 vm09 bash[21220]: audit 2026-03-07T07:02:02.491183+0000 mon.vm07 (mon.0) 902 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:02:03.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:03 vm07 bash[17031]: audit 2026-03-07T07:02:02.184841+0000 mon.vm07 (mon.0) 897 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:03.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:03 vm07 bash[17031]: audit 2026-03-07T07:02:02.184841+0000 mon.vm07 (mon.0) 897 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:03.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:03 vm07 bash[17031]: audit 2026-03-07T07:02:02.190104+0000 mon.vm07 (mon.0) 898 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:03.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:03 vm07 bash[17031]: audit 2026-03-07T07:02:02.190104+0000 mon.vm07 (mon.0) 898 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:03.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:03 vm07 bash[17031]: audit 2026-03-07T07:02:02.484574+0000 mon.vm07 (mon.0) 899 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:02:03.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:03 vm07 bash[17031]: audit 2026-03-07T07:02:02.484574+0000 mon.vm07 (mon.0) 899 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:02:03.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:03 vm07 bash[17031]: audit 2026-03-07T07:02:02.485136+0000 mon.vm07 (mon.0) 900 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:02:03.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:03 vm07 bash[17031]: audit 2026-03-07T07:02:02.485136+0000 mon.vm07 (mon.0) 900 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:02:03.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:03 vm07 bash[17031]: cluster 2026-03-07T07:02:02.486021+0000 mgr.vm07.yrfcuj (mgr.14201) 407 : cluster [DBG] pgmap v251: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:03.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:03 vm07 bash[17031]: cluster 2026-03-07T07:02:02.486021+0000 mgr.vm07.yrfcuj (mgr.14201) 407 : cluster [DBG] pgmap v251: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:03.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:03 vm07 bash[17031]: cluster 2026-03-07T07:02:02.486185+0000 mgr.vm07.yrfcuj (mgr.14201) 408 : cluster [DBG] pgmap v252: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:03.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:03 vm07 bash[17031]: cluster 2026-03-07T07:02:02.486185+0000 mgr.vm07.yrfcuj (mgr.14201) 408 : cluster [DBG] pgmap v252: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:03.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:03 vm07 bash[17031]: audit 2026-03-07T07:02:02.489897+0000 mon.vm07 (mon.0) 901 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:03.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:03 vm07 bash[17031]: audit 2026-03-07T07:02:02.489897+0000 mon.vm07 (mon.0) 901 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:02:03.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:03 vm07 bash[17031]: audit 2026-03-07T07:02:02.491183+0000 mon.vm07 (mon.0) 902 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:02:03.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:03 vm07 bash[17031]: audit 2026-03-07T07:02:02.491183+0000 mon.vm07 (mon.0) 902 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:02:04.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:04 vm09 bash[21220]: cluster 2026-03-07T07:02:04.486643+0000 mgr.vm07.yrfcuj (mgr.14201) 409 : cluster [DBG] pgmap v253: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:04.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:04 vm09 bash[21220]: cluster 2026-03-07T07:02:04.486643+0000 mgr.vm07.yrfcuj (mgr.14201) 409 : cluster [DBG] pgmap v253: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:04.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:04 vm07 bash[17031]: cluster 2026-03-07T07:02:04.486643+0000 mgr.vm07.yrfcuj (mgr.14201) 409 : cluster [DBG] pgmap v253: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:04.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:04 vm07 bash[17031]: cluster 2026-03-07T07:02:04.486643+0000 mgr.vm07.yrfcuj (mgr.14201) 409 : cluster [DBG] pgmap v253: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:05.603 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:02:05.754 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:02:05.754 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 3s ago 4m - - 2026-03-07T08:02:05.755 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 3s ago 4m 109M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:02:05.755 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 3s ago 4m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:02:05.755 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (4m) 3s ago 4m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:02:05.947 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:02:05.947 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:02:05.947 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:02:06.364 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:06 vm09 bash[21220]: audit 2026-03-07T07:02:05.947872+0000 mon.vm07 (mon.0) 903 : audit [DBG] from='client.? 192.168.123.107:0/46586737' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:06.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:06 vm09 bash[21220]: audit 2026-03-07T07:02:05.947872+0000 mon.vm07 (mon.0) 903 : audit [DBG] from='client.? 192.168.123.107:0/46586737' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:06.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:06 vm07 bash[17031]: audit 2026-03-07T07:02:05.947872+0000 mon.vm07 (mon.0) 903 : audit [DBG] from='client.? 192.168.123.107:0/46586737' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:06.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:06 vm07 bash[17031]: audit 2026-03-07T07:02:05.947872+0000 mon.vm07 (mon.0) 903 : audit [DBG] from='client.? 192.168.123.107:0/46586737' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:07.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:07 vm09 bash[21220]: audit 2026-03-07T07:02:05.590293+0000 mgr.vm07.yrfcuj (mgr.14201) 410 : audit [DBG] from='client.15120 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:07.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:07 vm09 bash[21220]: audit 2026-03-07T07:02:05.590293+0000 mgr.vm07.yrfcuj (mgr.14201) 410 : audit [DBG] from='client.15120 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:07.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:07 vm09 bash[21220]: audit 2026-03-07T07:02:05.752806+0000 mgr.vm07.yrfcuj (mgr.14201) 411 : audit [DBG] from='client.15124 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:07.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:07 vm09 bash[21220]: audit 2026-03-07T07:02:05.752806+0000 mgr.vm07.yrfcuj (mgr.14201) 411 : audit [DBG] from='client.15124 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:07.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:07 vm09 bash[21220]: cluster 2026-03-07T07:02:06.487033+0000 mgr.vm07.yrfcuj (mgr.14201) 412 : cluster [DBG] pgmap v254: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:07.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:07 vm09 bash[21220]: cluster 2026-03-07T07:02:06.487033+0000 mgr.vm07.yrfcuj (mgr.14201) 412 : cluster [DBG] pgmap v254: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:07.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:07 vm07 bash[17031]: audit 2026-03-07T07:02:05.590293+0000 mgr.vm07.yrfcuj (mgr.14201) 410 : audit [DBG] from='client.15120 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:07.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:07 vm07 bash[17031]: audit 2026-03-07T07:02:05.590293+0000 mgr.vm07.yrfcuj (mgr.14201) 410 : audit [DBG] from='client.15120 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:07.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:07 vm07 bash[17031]: audit 2026-03-07T07:02:05.752806+0000 mgr.vm07.yrfcuj (mgr.14201) 411 : audit [DBG] from='client.15124 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:07.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:07 vm07 bash[17031]: audit 2026-03-07T07:02:05.752806+0000 mgr.vm07.yrfcuj (mgr.14201) 411 : audit [DBG] from='client.15124 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:07.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:07 vm07 bash[17031]: cluster 2026-03-07T07:02:06.487033+0000 mgr.vm07.yrfcuj (mgr.14201) 412 : cluster [DBG] pgmap v254: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:07.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:07 vm07 bash[17031]: cluster 2026-03-07T07:02:06.487033+0000 mgr.vm07.yrfcuj (mgr.14201) 412 : cluster [DBG] pgmap v254: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:08.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:08 vm09 bash[21220]: cluster 2026-03-07T07:02:08.487490+0000 mgr.vm07.yrfcuj (mgr.14201) 413 : cluster [DBG] pgmap v255: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:08.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:08 vm09 bash[21220]: cluster 2026-03-07T07:02:08.487490+0000 mgr.vm07.yrfcuj (mgr.14201) 413 : cluster [DBG] pgmap v255: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:08.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:08 vm07 bash[17031]: cluster 2026-03-07T07:02:08.487490+0000 mgr.vm07.yrfcuj (mgr.14201) 413 : cluster [DBG] pgmap v255: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:08.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:08 vm07 bash[17031]: cluster 2026-03-07T07:02:08.487490+0000 mgr.vm07.yrfcuj (mgr.14201) 413 : cluster [DBG] pgmap v255: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:10 vm09 bash[21220]: audit 2026-03-07T07:02:10.131872+0000 mon.vm07 (mon.0) 904 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:02:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:10 vm09 bash[21220]: audit 2026-03-07T07:02:10.131872+0000 mon.vm07 (mon.0) 904 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:02:10.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:10 vm07 bash[17031]: audit 2026-03-07T07:02:10.131872+0000 mon.vm07 (mon.0) 904 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:02:10.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:10 vm07 bash[17031]: audit 2026-03-07T07:02:10.131872+0000 mon.vm07 (mon.0) 904 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:02:11.122 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:02:11.283 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:02:11.283 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 9s ago 4m - - 2026-03-07T08:02:11.283 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 9s ago 4m 109M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:02:11.283 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 9s ago 4m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:02:11.283 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (4m) 9s ago 4m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:02:11.465 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:02:11.465 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:02:11.465 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:02:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:11 vm09 bash[21220]: cluster 2026-03-07T07:02:10.487878+0000 mgr.vm07.yrfcuj (mgr.14201) 414 : cluster [DBG] pgmap v256: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:11 vm09 bash[21220]: cluster 2026-03-07T07:02:10.487878+0000 mgr.vm07.yrfcuj (mgr.14201) 414 : cluster [DBG] pgmap v256: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:11.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:11 vm07 bash[17031]: cluster 2026-03-07T07:02:10.487878+0000 mgr.vm07.yrfcuj (mgr.14201) 414 : cluster [DBG] pgmap v256: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:11.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:11 vm07 bash[17031]: cluster 2026-03-07T07:02:10.487878+0000 mgr.vm07.yrfcuj (mgr.14201) 414 : cluster [DBG] pgmap v256: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:12.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:12 vm09 bash[21220]: audit 2026-03-07T07:02:11.108559+0000 mgr.vm07.yrfcuj (mgr.14201) 415 : audit [DBG] from='client.15132 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:12.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:12 vm09 bash[21220]: audit 2026-03-07T07:02:11.108559+0000 mgr.vm07.yrfcuj (mgr.14201) 415 : audit [DBG] from='client.15132 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:12.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:12 vm09 bash[21220]: audit 2026-03-07T07:02:11.281252+0000 mgr.vm07.yrfcuj (mgr.14201) 416 : audit [DBG] from='client.24733 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:12.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:12 vm09 bash[21220]: audit 2026-03-07T07:02:11.281252+0000 mgr.vm07.yrfcuj (mgr.14201) 416 : audit [DBG] from='client.24733 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:12.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:12 vm09 bash[21220]: audit 2026-03-07T07:02:11.465507+0000 mon.vm07 (mon.0) 905 : audit [DBG] from='client.? 192.168.123.107:0/2422448946' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:12.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:12 vm09 bash[21220]: audit 2026-03-07T07:02:11.465507+0000 mon.vm07 (mon.0) 905 : audit [DBG] from='client.? 192.168.123.107:0/2422448946' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:12.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:12 vm07 bash[17031]: audit 2026-03-07T07:02:11.108559+0000 mgr.vm07.yrfcuj (mgr.14201) 415 : audit [DBG] from='client.15132 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:12.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:12 vm07 bash[17031]: audit 2026-03-07T07:02:11.108559+0000 mgr.vm07.yrfcuj (mgr.14201) 415 : audit [DBG] from='client.15132 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:12.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:12 vm07 bash[17031]: audit 2026-03-07T07:02:11.281252+0000 mgr.vm07.yrfcuj (mgr.14201) 416 : audit [DBG] from='client.24733 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:12.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:12 vm07 bash[17031]: audit 2026-03-07T07:02:11.281252+0000 mgr.vm07.yrfcuj (mgr.14201) 416 : audit [DBG] from='client.24733 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:12.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:12 vm07 bash[17031]: audit 2026-03-07T07:02:11.465507+0000 mon.vm07 (mon.0) 905 : audit [DBG] from='client.? 192.168.123.107:0/2422448946' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:12.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:12 vm07 bash[17031]: audit 2026-03-07T07:02:11.465507+0000 mon.vm07 (mon.0) 905 : audit [DBG] from='client.? 192.168.123.107:0/2422448946' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:13 vm09 bash[21220]: cluster 2026-03-07T07:02:12.488292+0000 mgr.vm07.yrfcuj (mgr.14201) 417 : cluster [DBG] pgmap v257: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:13 vm09 bash[21220]: cluster 2026-03-07T07:02:12.488292+0000 mgr.vm07.yrfcuj (mgr.14201) 417 : cluster [DBG] pgmap v257: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:13.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:13 vm07 bash[17031]: cluster 2026-03-07T07:02:12.488292+0000 mgr.vm07.yrfcuj (mgr.14201) 417 : cluster [DBG] pgmap v257: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:13.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:13 vm07 bash[17031]: cluster 2026-03-07T07:02:12.488292+0000 mgr.vm07.yrfcuj (mgr.14201) 417 : cluster [DBG] pgmap v257: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:14.844 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:14 vm07 bash[17031]: cluster 2026-03-07T07:02:14.488726+0000 mgr.vm07.yrfcuj (mgr.14201) 418 : cluster [DBG] pgmap v258: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:14.844 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:14 vm07 bash[17031]: cluster 2026-03-07T07:02:14.488726+0000 mgr.vm07.yrfcuj (mgr.14201) 418 : cluster [DBG] pgmap v258: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:14.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:14 vm09 bash[21220]: cluster 2026-03-07T07:02:14.488726+0000 mgr.vm07.yrfcuj (mgr.14201) 418 : cluster [DBG] pgmap v258: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:14.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:14 vm09 bash[21220]: cluster 2026-03-07T07:02:14.488726+0000 mgr.vm07.yrfcuj (mgr.14201) 418 : cluster [DBG] pgmap v258: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:16.646 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:02:16.805 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:02:16.805 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 14s ago 4m - - 2026-03-07T08:02:16.805 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 14s ago 4m 109M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:02:16.805 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 14s ago 4m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:02:16.805 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (4m) 14s ago 4m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:02:16.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:16 vm09 bash[21220]: cluster 2026-03-07T07:02:16.489083+0000 mgr.vm07.yrfcuj (mgr.14201) 419 : cluster [DBG] pgmap v259: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:16.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:16 vm09 bash[21220]: cluster 2026-03-07T07:02:16.489083+0000 mgr.vm07.yrfcuj (mgr.14201) 419 : cluster [DBG] pgmap v259: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:16.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:16 vm07 bash[17031]: cluster 2026-03-07T07:02:16.489083+0000 mgr.vm07.yrfcuj (mgr.14201) 419 : cluster [DBG] pgmap v259: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:16.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:16 vm07 bash[17031]: cluster 2026-03-07T07:02:16.489083+0000 mgr.vm07.yrfcuj (mgr.14201) 419 : cluster [DBG] pgmap v259: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:16.985 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:02:16.985 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:02:16.985 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:02:17.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:17 vm09 bash[21220]: audit 2026-03-07T07:02:16.632055+0000 mgr.vm07.yrfcuj (mgr.14201) 420 : audit [DBG] from='client.15144 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:17.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:17 vm09 bash[21220]: audit 2026-03-07T07:02:16.632055+0000 mgr.vm07.yrfcuj (mgr.14201) 420 : audit [DBG] from='client.15144 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:17.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:17 vm09 bash[21220]: audit 2026-03-07T07:02:16.803170+0000 mgr.vm07.yrfcuj (mgr.14201) 421 : audit [DBG] from='client.15148 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:17.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:17 vm09 bash[21220]: audit 2026-03-07T07:02:16.803170+0000 mgr.vm07.yrfcuj (mgr.14201) 421 : audit [DBG] from='client.15148 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:17.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:17 vm09 bash[21220]: audit 2026-03-07T07:02:16.984447+0000 mon.vm09 (mon.1) 33 : audit [DBG] from='client.? 192.168.123.107:0/2118138255' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:17.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:17 vm09 bash[21220]: audit 2026-03-07T07:02:16.984447+0000 mon.vm09 (mon.1) 33 : audit [DBG] from='client.? 192.168.123.107:0/2118138255' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:17.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:17 vm07 bash[17031]: audit 2026-03-07T07:02:16.632055+0000 mgr.vm07.yrfcuj (mgr.14201) 420 : audit [DBG] from='client.15144 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:17.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:17 vm07 bash[17031]: audit 2026-03-07T07:02:16.632055+0000 mgr.vm07.yrfcuj (mgr.14201) 420 : audit [DBG] from='client.15144 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:17.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:17 vm07 bash[17031]: audit 2026-03-07T07:02:16.803170+0000 mgr.vm07.yrfcuj (mgr.14201) 421 : audit [DBG] from='client.15148 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:17.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:17 vm07 bash[17031]: audit 2026-03-07T07:02:16.803170+0000 mgr.vm07.yrfcuj (mgr.14201) 421 : audit [DBG] from='client.15148 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:17.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:17 vm07 bash[17031]: audit 2026-03-07T07:02:16.984447+0000 mon.vm09 (mon.1) 33 : audit [DBG] from='client.? 192.168.123.107:0/2118138255' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:17.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:17 vm07 bash[17031]: audit 2026-03-07T07:02:16.984447+0000 mon.vm09 (mon.1) 33 : audit [DBG] from='client.? 192.168.123.107:0/2118138255' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:18 vm09 bash[21220]: cluster 2026-03-07T07:02:18.489584+0000 mgr.vm07.yrfcuj (mgr.14201) 422 : cluster [DBG] pgmap v260: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:18 vm09 bash[21220]: cluster 2026-03-07T07:02:18.489584+0000 mgr.vm07.yrfcuj (mgr.14201) 422 : cluster [DBG] pgmap v260: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:18.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:18 vm07 bash[17031]: cluster 2026-03-07T07:02:18.489584+0000 mgr.vm07.yrfcuj (mgr.14201) 422 : cluster [DBG] pgmap v260: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:18.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:18 vm07 bash[17031]: cluster 2026-03-07T07:02:18.489584+0000 mgr.vm07.yrfcuj (mgr.14201) 422 : cluster [DBG] pgmap v260: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:20 vm09 bash[21220]: cluster 2026-03-07T07:02:20.489998+0000 mgr.vm07.yrfcuj (mgr.14201) 423 : cluster [DBG] pgmap v261: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:20 vm09 bash[21220]: cluster 2026-03-07T07:02:20.489998+0000 mgr.vm07.yrfcuj (mgr.14201) 423 : cluster [DBG] pgmap v261: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:20.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:20 vm07 bash[17031]: cluster 2026-03-07T07:02:20.489998+0000 mgr.vm07.yrfcuj (mgr.14201) 423 : cluster [DBG] pgmap v261: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:20.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:20 vm07 bash[17031]: cluster 2026-03-07T07:02:20.489998+0000 mgr.vm07.yrfcuj (mgr.14201) 423 : cluster [DBG] pgmap v261: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:22.160 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:02:22.308 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:02:22.308 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 20s ago 4m - - 2026-03-07T08:02:22.308 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 20s ago 4m 109M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:02:22.308 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 20s ago 4m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:02:22.308 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (4m) 20s ago 4m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:02:22.485 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:02:22.486 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:02:22.486 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:02:22.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:22 vm09 bash[21220]: audit 2026-03-07T07:02:22.485998+0000 mon.vm07 (mon.0) 906 : audit [DBG] from='client.? 192.168.123.107:0/1372157924' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:22.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:22 vm09 bash[21220]: audit 2026-03-07T07:02:22.485998+0000 mon.vm07 (mon.0) 906 : audit [DBG] from='client.? 192.168.123.107:0/1372157924' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:22.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:22 vm07 bash[17031]: audit 2026-03-07T07:02:22.485998+0000 mon.vm07 (mon.0) 906 : audit [DBG] from='client.? 192.168.123.107:0/1372157924' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:22.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:22 vm07 bash[17031]: audit 2026-03-07T07:02:22.485998+0000 mon.vm07 (mon.0) 906 : audit [DBG] from='client.? 192.168.123.107:0/1372157924' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:23.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:23 vm09 bash[21220]: audit 2026-03-07T07:02:22.146235+0000 mgr.vm07.yrfcuj (mgr.14201) 424 : audit [DBG] from='client.15156 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:23.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:23 vm09 bash[21220]: audit 2026-03-07T07:02:22.146235+0000 mgr.vm07.yrfcuj (mgr.14201) 424 : audit [DBG] from='client.15156 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:23.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:23 vm09 bash[21220]: audit 2026-03-07T07:02:22.305876+0000 mgr.vm07.yrfcuj (mgr.14201) 425 : audit [DBG] from='client.15160 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:23.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:23 vm09 bash[21220]: audit 2026-03-07T07:02:22.305876+0000 mgr.vm07.yrfcuj (mgr.14201) 425 : audit [DBG] from='client.15160 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:23.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:23 vm09 bash[21220]: cluster 2026-03-07T07:02:22.490416+0000 mgr.vm07.yrfcuj (mgr.14201) 426 : cluster [DBG] pgmap v262: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:23.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:23 vm09 bash[21220]: cluster 2026-03-07T07:02:22.490416+0000 mgr.vm07.yrfcuj (mgr.14201) 426 : cluster [DBG] pgmap v262: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:23.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:23 vm07 bash[17031]: audit 2026-03-07T07:02:22.146235+0000 mgr.vm07.yrfcuj (mgr.14201) 424 : audit [DBG] from='client.15156 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:23.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:23 vm07 bash[17031]: audit 2026-03-07T07:02:22.146235+0000 mgr.vm07.yrfcuj (mgr.14201) 424 : audit [DBG] from='client.15156 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:23.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:23 vm07 bash[17031]: audit 2026-03-07T07:02:22.305876+0000 mgr.vm07.yrfcuj (mgr.14201) 425 : audit [DBG] from='client.15160 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:23.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:23 vm07 bash[17031]: audit 2026-03-07T07:02:22.305876+0000 mgr.vm07.yrfcuj (mgr.14201) 425 : audit [DBG] from='client.15160 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:23.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:23 vm07 bash[17031]: cluster 2026-03-07T07:02:22.490416+0000 mgr.vm07.yrfcuj (mgr.14201) 426 : cluster [DBG] pgmap v262: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:23.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:23 vm07 bash[17031]: cluster 2026-03-07T07:02:22.490416+0000 mgr.vm07.yrfcuj (mgr.14201) 426 : cluster [DBG] pgmap v262: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:24.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:24 vm09 bash[21220]: cluster 2026-03-07T07:02:24.490918+0000 mgr.vm07.yrfcuj (mgr.14201) 427 : cluster [DBG] pgmap v263: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:24.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:24 vm09 bash[21220]: cluster 2026-03-07T07:02:24.490918+0000 mgr.vm07.yrfcuj (mgr.14201) 427 : cluster [DBG] pgmap v263: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:24.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:24 vm07 bash[17031]: cluster 2026-03-07T07:02:24.490918+0000 mgr.vm07.yrfcuj (mgr.14201) 427 : cluster [DBG] pgmap v263: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:24.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:24 vm07 bash[17031]: cluster 2026-03-07T07:02:24.490918+0000 mgr.vm07.yrfcuj (mgr.14201) 427 : cluster [DBG] pgmap v263: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:25 vm09 bash[21220]: audit 2026-03-07T07:02:25.131852+0000 mon.vm07 (mon.0) 907 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:02:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:25 vm09 bash[21220]: audit 2026-03-07T07:02:25.131852+0000 mon.vm07 (mon.0) 907 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:02:25.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:25 vm07 bash[17031]: audit 2026-03-07T07:02:25.131852+0000 mon.vm07 (mon.0) 907 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:02:25.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:25 vm07 bash[17031]: audit 2026-03-07T07:02:25.131852+0000 mon.vm07 (mon.0) 907 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:02:26.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:26 vm09 bash[21220]: cluster 2026-03-07T07:02:26.491353+0000 mgr.vm07.yrfcuj (mgr.14201) 428 : cluster [DBG] pgmap v264: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:26.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:26 vm09 bash[21220]: cluster 2026-03-07T07:02:26.491353+0000 mgr.vm07.yrfcuj (mgr.14201) 428 : cluster [DBG] pgmap v264: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:26.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:26 vm07 bash[17031]: cluster 2026-03-07T07:02:26.491353+0000 mgr.vm07.yrfcuj (mgr.14201) 428 : cluster [DBG] pgmap v264: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:26.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:26 vm07 bash[17031]: cluster 2026-03-07T07:02:26.491353+0000 mgr.vm07.yrfcuj (mgr.14201) 428 : cluster [DBG] pgmap v264: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:27.660 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:02:27.812 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:02:27.812 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 25s ago 4m - - 2026-03-07T08:02:27.812 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 25s ago 4m 109M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:02:27.812 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 25s ago 4m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:02:27.812 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (4m) 25s ago 4m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:02:28.003 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:02:28.004 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:02:28.004 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:02:28.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:28 vm09 bash[21220]: audit 2026-03-07T07:02:28.003106+0000 mon.vm07 (mon.0) 908 : audit [DBG] from='client.? 192.168.123.107:0/2701658251' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:28.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:28 vm09 bash[21220]: audit 2026-03-07T07:02:28.003106+0000 mon.vm07 (mon.0) 908 : audit [DBG] from='client.? 192.168.123.107:0/2701658251' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:28.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:28 vm07 bash[17031]: audit 2026-03-07T07:02:28.003106+0000 mon.vm07 (mon.0) 908 : audit [DBG] from='client.? 192.168.123.107:0/2701658251' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:28.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:28 vm07 bash[17031]: audit 2026-03-07T07:02:28.003106+0000 mon.vm07 (mon.0) 908 : audit [DBG] from='client.? 192.168.123.107:0/2701658251' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:29.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:29 vm09 bash[21220]: audit 2026-03-07T07:02:27.642958+0000 mgr.vm07.yrfcuj (mgr.14201) 429 : audit [DBG] from='client.15168 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:29.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:29 vm09 bash[21220]: audit 2026-03-07T07:02:27.642958+0000 mgr.vm07.yrfcuj (mgr.14201) 429 : audit [DBG] from='client.15168 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:29.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:29 vm09 bash[21220]: audit 2026-03-07T07:02:27.810072+0000 mgr.vm07.yrfcuj (mgr.14201) 430 : audit [DBG] from='client.15172 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:29.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:29 vm09 bash[21220]: audit 2026-03-07T07:02:27.810072+0000 mgr.vm07.yrfcuj (mgr.14201) 430 : audit [DBG] from='client.15172 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:29.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:29 vm09 bash[21220]: cluster 2026-03-07T07:02:28.491817+0000 mgr.vm07.yrfcuj (mgr.14201) 431 : cluster [DBG] pgmap v265: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:29.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:29 vm09 bash[21220]: cluster 2026-03-07T07:02:28.491817+0000 mgr.vm07.yrfcuj (mgr.14201) 431 : cluster [DBG] pgmap v265: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:29.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:29 vm07 bash[17031]: audit 2026-03-07T07:02:27.642958+0000 mgr.vm07.yrfcuj (mgr.14201) 429 : audit [DBG] from='client.15168 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:29.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:29 vm07 bash[17031]: audit 2026-03-07T07:02:27.642958+0000 mgr.vm07.yrfcuj (mgr.14201) 429 : audit [DBG] from='client.15168 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:29.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:29 vm07 bash[17031]: audit 2026-03-07T07:02:27.810072+0000 mgr.vm07.yrfcuj (mgr.14201) 430 : audit [DBG] from='client.15172 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:29.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:29 vm07 bash[17031]: audit 2026-03-07T07:02:27.810072+0000 mgr.vm07.yrfcuj (mgr.14201) 430 : audit [DBG] from='client.15172 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:29.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:29 vm07 bash[17031]: cluster 2026-03-07T07:02:28.491817+0000 mgr.vm07.yrfcuj (mgr.14201) 431 : cluster [DBG] pgmap v265: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:29.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:29 vm07 bash[17031]: cluster 2026-03-07T07:02:28.491817+0000 mgr.vm07.yrfcuj (mgr.14201) 431 : cluster [DBG] pgmap v265: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:30.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:30 vm09 bash[21220]: cluster 2026-03-07T07:02:30.492217+0000 mgr.vm07.yrfcuj (mgr.14201) 432 : cluster [DBG] pgmap v266: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:30.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:30 vm09 bash[21220]: cluster 2026-03-07T07:02:30.492217+0000 mgr.vm07.yrfcuj (mgr.14201) 432 : cluster [DBG] pgmap v266: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:30.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:30 vm07 bash[17031]: cluster 2026-03-07T07:02:30.492217+0000 mgr.vm07.yrfcuj (mgr.14201) 432 : cluster [DBG] pgmap v266: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:30.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:30 vm07 bash[17031]: cluster 2026-03-07T07:02:30.492217+0000 mgr.vm07.yrfcuj (mgr.14201) 432 : cluster [DBG] pgmap v266: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:32.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:32 vm09 bash[21220]: cluster 2026-03-07T07:02:32.492702+0000 mgr.vm07.yrfcuj (mgr.14201) 433 : cluster [DBG] pgmap v267: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:32.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:32 vm09 bash[21220]: cluster 2026-03-07T07:02:32.492702+0000 mgr.vm07.yrfcuj (mgr.14201) 433 : cluster [DBG] pgmap v267: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:32.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:32 vm07 bash[17031]: cluster 2026-03-07T07:02:32.492702+0000 mgr.vm07.yrfcuj (mgr.14201) 433 : cluster [DBG] pgmap v267: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:32.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:32 vm07 bash[17031]: cluster 2026-03-07T07:02:32.492702+0000 mgr.vm07.yrfcuj (mgr.14201) 433 : cluster [DBG] pgmap v267: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:33.178 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:02:33.340 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:02:33.340 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 31s ago 4m - - 2026-03-07T08:02:33.340 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 31s ago 4m 109M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:02:33.340 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 31s ago 4m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:02:33.340 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (4m) 31s ago 4m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:02:33.531 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:02:33.531 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:02:33.531 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:02:33.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:33 vm09 bash[21220]: audit 2026-03-07T07:02:33.163504+0000 mgr.vm07.yrfcuj (mgr.14201) 434 : audit [DBG] from='client.15180 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:33.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:33 vm09 bash[21220]: audit 2026-03-07T07:02:33.163504+0000 mgr.vm07.yrfcuj (mgr.14201) 434 : audit [DBG] from='client.15180 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:33.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:33 vm09 bash[21220]: audit 2026-03-07T07:02:33.337930+0000 mgr.vm07.yrfcuj (mgr.14201) 435 : audit [DBG] from='client.15184 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:33.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:33 vm09 bash[21220]: audit 2026-03-07T07:02:33.337930+0000 mgr.vm07.yrfcuj (mgr.14201) 435 : audit [DBG] from='client.15184 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:33.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:33 vm09 bash[21220]: audit 2026-03-07T07:02:33.531343+0000 mon.vm07 (mon.0) 909 : audit [DBG] from='client.? 192.168.123.107:0/2111875947' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:33.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:33 vm09 bash[21220]: audit 2026-03-07T07:02:33.531343+0000 mon.vm07 (mon.0) 909 : audit [DBG] from='client.? 192.168.123.107:0/2111875947' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:33.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:33 vm07 bash[17031]: audit 2026-03-07T07:02:33.163504+0000 mgr.vm07.yrfcuj (mgr.14201) 434 : audit [DBG] from='client.15180 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:33.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:33 vm07 bash[17031]: audit 2026-03-07T07:02:33.163504+0000 mgr.vm07.yrfcuj (mgr.14201) 434 : audit [DBG] from='client.15180 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:33.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:33 vm07 bash[17031]: audit 2026-03-07T07:02:33.337930+0000 mgr.vm07.yrfcuj (mgr.14201) 435 : audit [DBG] from='client.15184 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:33.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:33 vm07 bash[17031]: audit 2026-03-07T07:02:33.337930+0000 mgr.vm07.yrfcuj (mgr.14201) 435 : audit [DBG] from='client.15184 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:33.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:33 vm07 bash[17031]: audit 2026-03-07T07:02:33.531343+0000 mon.vm07 (mon.0) 909 : audit [DBG] from='client.? 192.168.123.107:0/2111875947' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:33.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:33 vm07 bash[17031]: audit 2026-03-07T07:02:33.531343+0000 mon.vm07 (mon.0) 909 : audit [DBG] from='client.? 192.168.123.107:0/2111875947' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:34.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:34 vm09 bash[21220]: cluster 2026-03-07T07:02:34.493176+0000 mgr.vm07.yrfcuj (mgr.14201) 436 : cluster [DBG] pgmap v268: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:34.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:34 vm09 bash[21220]: cluster 2026-03-07T07:02:34.493176+0000 mgr.vm07.yrfcuj (mgr.14201) 436 : cluster [DBG] pgmap v268: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:34.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:34 vm07 bash[17031]: cluster 2026-03-07T07:02:34.493176+0000 mgr.vm07.yrfcuj (mgr.14201) 436 : cluster [DBG] pgmap v268: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:34.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:34 vm07 bash[17031]: cluster 2026-03-07T07:02:34.493176+0000 mgr.vm07.yrfcuj (mgr.14201) 436 : cluster [DBG] pgmap v268: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:36 vm09 bash[21220]: cluster 2026-03-07T07:02:36.493589+0000 mgr.vm07.yrfcuj (mgr.14201) 437 : cluster [DBG] pgmap v269: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:36 vm09 bash[21220]: cluster 2026-03-07T07:02:36.493589+0000 mgr.vm07.yrfcuj (mgr.14201) 437 : cluster [DBG] pgmap v269: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:36 vm07 bash[17031]: cluster 2026-03-07T07:02:36.493589+0000 mgr.vm07.yrfcuj (mgr.14201) 437 : cluster [DBG] pgmap v269: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:36 vm07 bash[17031]: cluster 2026-03-07T07:02:36.493589+0000 mgr.vm07.yrfcuj (mgr.14201) 437 : cluster [DBG] pgmap v269: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:38.710 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:02:38.859 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:02:38.859 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 36s ago 5m - - 2026-03-07T08:02:38.859 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 36s ago 4m 109M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:02:38.859 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (5m) 36s ago 5m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:02:38.859 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (5m) 36s ago 5m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:02:38.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:38 vm09 bash[21220]: cluster 2026-03-07T07:02:38.494145+0000 mgr.vm07.yrfcuj (mgr.14201) 438 : cluster [DBG] pgmap v270: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:38.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:38 vm09 bash[21220]: cluster 2026-03-07T07:02:38.494145+0000 mgr.vm07.yrfcuj (mgr.14201) 438 : cluster [DBG] pgmap v270: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:38.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:38 vm07 bash[17031]: cluster 2026-03-07T07:02:38.494145+0000 mgr.vm07.yrfcuj (mgr.14201) 438 : cluster [DBG] pgmap v270: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:38.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:38 vm07 bash[17031]: cluster 2026-03-07T07:02:38.494145+0000 mgr.vm07.yrfcuj (mgr.14201) 438 : cluster [DBG] pgmap v270: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:39.055 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:02:39.055 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:02:39.055 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:02:39.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:39 vm09 bash[21220]: audit 2026-03-07T07:02:38.696122+0000 mgr.vm07.yrfcuj (mgr.14201) 439 : audit [DBG] from='client.15192 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:39.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:39 vm09 bash[21220]: audit 2026-03-07T07:02:38.696122+0000 mgr.vm07.yrfcuj (mgr.14201) 439 : audit [DBG] from='client.15192 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:39.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:39 vm09 bash[21220]: audit 2026-03-07T07:02:38.856221+0000 mgr.vm07.yrfcuj (mgr.14201) 440 : audit [DBG] from='client.15196 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:39.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:39 vm09 bash[21220]: audit 2026-03-07T07:02:38.856221+0000 mgr.vm07.yrfcuj (mgr.14201) 440 : audit [DBG] from='client.15196 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:39.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:39 vm09 bash[21220]: audit 2026-03-07T07:02:39.055387+0000 mon.vm07 (mon.0) 910 : audit [DBG] from='client.? 192.168.123.107:0/3770751096' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:39.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:39 vm09 bash[21220]: audit 2026-03-07T07:02:39.055387+0000 mon.vm07 (mon.0) 910 : audit [DBG] from='client.? 192.168.123.107:0/3770751096' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:39.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:39 vm07 bash[17031]: audit 2026-03-07T07:02:38.696122+0000 mgr.vm07.yrfcuj (mgr.14201) 439 : audit [DBG] from='client.15192 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:39.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:39 vm07 bash[17031]: audit 2026-03-07T07:02:38.696122+0000 mgr.vm07.yrfcuj (mgr.14201) 439 : audit [DBG] from='client.15192 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:39.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:39 vm07 bash[17031]: audit 2026-03-07T07:02:38.856221+0000 mgr.vm07.yrfcuj (mgr.14201) 440 : audit [DBG] from='client.15196 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:39.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:39 vm07 bash[17031]: audit 2026-03-07T07:02:38.856221+0000 mgr.vm07.yrfcuj (mgr.14201) 440 : audit [DBG] from='client.15196 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:39.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:39 vm07 bash[17031]: audit 2026-03-07T07:02:39.055387+0000 mon.vm07 (mon.0) 910 : audit [DBG] from='client.? 192.168.123.107:0/3770751096' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:39.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:39 vm07 bash[17031]: audit 2026-03-07T07:02:39.055387+0000 mon.vm07 (mon.0) 910 : audit [DBG] from='client.? 192.168.123.107:0/3770751096' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:40.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:40 vm07 bash[17031]: audit 2026-03-07T07:02:40.132641+0000 mon.vm07 (mon.0) 911 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:02:40.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:40 vm07 bash[17031]: audit 2026-03-07T07:02:40.132641+0000 mon.vm07 (mon.0) 911 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:02:40.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:40 vm07 bash[17031]: cluster 2026-03-07T07:02:40.494584+0000 mgr.vm07.yrfcuj (mgr.14201) 441 : cluster [DBG] pgmap v271: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:40.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:40 vm07 bash[17031]: cluster 2026-03-07T07:02:40.494584+0000 mgr.vm07.yrfcuj (mgr.14201) 441 : cluster [DBG] pgmap v271: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:41.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:40 vm09 bash[21220]: audit 2026-03-07T07:02:40.132641+0000 mon.vm07 (mon.0) 911 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:02:41.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:40 vm09 bash[21220]: audit 2026-03-07T07:02:40.132641+0000 mon.vm07 (mon.0) 911 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:02:41.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:40 vm09 bash[21220]: cluster 2026-03-07T07:02:40.494584+0000 mgr.vm07.yrfcuj (mgr.14201) 441 : cluster [DBG] pgmap v271: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:41.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:40 vm09 bash[21220]: cluster 2026-03-07T07:02:40.494584+0000 mgr.vm07.yrfcuj (mgr.14201) 441 : cluster [DBG] pgmap v271: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:43.083 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:42 vm09 bash[21220]: cluster 2026-03-07T07:02:42.494994+0000 mgr.vm07.yrfcuj (mgr.14201) 442 : cluster [DBG] pgmap v272: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:43.084 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:42 vm09 bash[21220]: cluster 2026-03-07T07:02:42.494994+0000 mgr.vm07.yrfcuj (mgr.14201) 442 : cluster [DBG] pgmap v272: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:43.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:42 vm07 bash[17031]: cluster 2026-03-07T07:02:42.494994+0000 mgr.vm07.yrfcuj (mgr.14201) 442 : cluster [DBG] pgmap v272: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:43.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:42 vm07 bash[17031]: cluster 2026-03-07T07:02:42.494994+0000 mgr.vm07.yrfcuj (mgr.14201) 442 : cluster [DBG] pgmap v272: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:44.230 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:02:44.388 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:02:44.388 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 42s ago 5m - - 2026-03-07T08:02:44.388 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 42s ago 5m 109M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:02:44.388 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (5m) 42s ago 5m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:02:44.388 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (5m) 42s ago 5m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:02:44.581 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:02:44.581 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:02:44.581 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:02:44.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:44 vm09 bash[21220]: audit 2026-03-07T07:02:44.215102+0000 mgr.vm07.yrfcuj (mgr.14201) 443 : audit [DBG] from='client.15204 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:44.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:44 vm09 bash[21220]: audit 2026-03-07T07:02:44.215102+0000 mgr.vm07.yrfcuj (mgr.14201) 443 : audit [DBG] from='client.15204 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:44.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:44 vm09 bash[21220]: audit 2026-03-07T07:02:44.385737+0000 mgr.vm07.yrfcuj (mgr.14201) 444 : audit [DBG] from='client.15208 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:44.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:44 vm09 bash[21220]: audit 2026-03-07T07:02:44.385737+0000 mgr.vm07.yrfcuj (mgr.14201) 444 : audit [DBG] from='client.15208 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:44.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:44 vm09 bash[21220]: cluster 2026-03-07T07:02:44.495361+0000 mgr.vm07.yrfcuj (mgr.14201) 445 : cluster [DBG] pgmap v273: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:44.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:44 vm09 bash[21220]: cluster 2026-03-07T07:02:44.495361+0000 mgr.vm07.yrfcuj (mgr.14201) 445 : cluster [DBG] pgmap v273: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:44.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:44 vm09 bash[21220]: audit 2026-03-07T07:02:44.580627+0000 mon.vm09 (mon.1) 34 : audit [DBG] from='client.? 192.168.123.107:0/3814880164' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:44.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:44 vm09 bash[21220]: audit 2026-03-07T07:02:44.580627+0000 mon.vm09 (mon.1) 34 : audit [DBG] from='client.? 192.168.123.107:0/3814880164' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:44.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:44 vm07 bash[17031]: audit 2026-03-07T07:02:44.215102+0000 mgr.vm07.yrfcuj (mgr.14201) 443 : audit [DBG] from='client.15204 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:44.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:44 vm07 bash[17031]: audit 2026-03-07T07:02:44.215102+0000 mgr.vm07.yrfcuj (mgr.14201) 443 : audit [DBG] from='client.15204 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:44.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:44 vm07 bash[17031]: audit 2026-03-07T07:02:44.385737+0000 mgr.vm07.yrfcuj (mgr.14201) 444 : audit [DBG] from='client.15208 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:44.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:44 vm07 bash[17031]: audit 2026-03-07T07:02:44.385737+0000 mgr.vm07.yrfcuj (mgr.14201) 444 : audit [DBG] from='client.15208 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:44.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:44 vm07 bash[17031]: cluster 2026-03-07T07:02:44.495361+0000 mgr.vm07.yrfcuj (mgr.14201) 445 : cluster [DBG] pgmap v273: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:44.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:44 vm07 bash[17031]: cluster 2026-03-07T07:02:44.495361+0000 mgr.vm07.yrfcuj (mgr.14201) 445 : cluster [DBG] pgmap v273: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:44.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:44 vm07 bash[17031]: audit 2026-03-07T07:02:44.580627+0000 mon.vm09 (mon.1) 34 : audit [DBG] from='client.? 192.168.123.107:0/3814880164' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:44.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:44 vm07 bash[17031]: audit 2026-03-07T07:02:44.580627+0000 mon.vm09 (mon.1) 34 : audit [DBG] from='client.? 192.168.123.107:0/3814880164' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:46.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:46 vm09 bash[21220]: cluster 2026-03-07T07:02:46.495675+0000 mgr.vm07.yrfcuj (mgr.14201) 446 : cluster [DBG] pgmap v274: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:46.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:46 vm09 bash[21220]: cluster 2026-03-07T07:02:46.495675+0000 mgr.vm07.yrfcuj (mgr.14201) 446 : cluster [DBG] pgmap v274: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:46.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:46 vm07 bash[17031]: cluster 2026-03-07T07:02:46.495675+0000 mgr.vm07.yrfcuj (mgr.14201) 446 : cluster [DBG] pgmap v274: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:46.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:46 vm07 bash[17031]: cluster 2026-03-07T07:02:46.495675+0000 mgr.vm07.yrfcuj (mgr.14201) 446 : cluster [DBG] pgmap v274: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:02:48.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:48 vm09 bash[21220]: cluster 2026-03-07T07:02:48.496072+0000 mgr.vm07.yrfcuj (mgr.14201) 447 : cluster [DBG] pgmap v275: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:48 vm09 bash[21220]: cluster 2026-03-07T07:02:48.496072+0000 mgr.vm07.yrfcuj (mgr.14201) 447 : cluster [DBG] pgmap v275: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:48.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:48 vm07 bash[17031]: cluster 2026-03-07T07:02:48.496072+0000 mgr.vm07.yrfcuj (mgr.14201) 447 : cluster [DBG] pgmap v275: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:48.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:48 vm07 bash[17031]: cluster 2026-03-07T07:02:48.496072+0000 mgr.vm07.yrfcuj (mgr.14201) 447 : cluster [DBG] pgmap v275: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:49.758 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:02:49.914 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:02:49.914 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 47s ago 5m - - 2026-03-07T08:02:49.914 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 47s ago 5m 109M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:02:49.914 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (5m) 47s ago 5m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:02:49.914 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (5m) 47s ago 5m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:02:50.111 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:02:50.111 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:02:50.111 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:02:50.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:50 vm09 bash[21220]: audit 2026-03-07T07:02:50.110947+0000 mon.vm07 (mon.0) 912 : audit [DBG] from='client.? 192.168.123.107:0/1242702276' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:50.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:50 vm09 bash[21220]: audit 2026-03-07T07:02:50.110947+0000 mon.vm07 (mon.0) 912 : audit [DBG] from='client.? 192.168.123.107:0/1242702276' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:50.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:50 vm07 bash[17031]: audit 2026-03-07T07:02:50.110947+0000 mon.vm07 (mon.0) 912 : audit [DBG] from='client.? 192.168.123.107:0/1242702276' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:50.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:50 vm07 bash[17031]: audit 2026-03-07T07:02:50.110947+0000 mon.vm07 (mon.0) 912 : audit [DBG] from='client.? 192.168.123.107:0/1242702276' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:51.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:51 vm09 bash[21220]: audit 2026-03-07T07:02:49.744139+0000 mgr.vm07.yrfcuj (mgr.14201) 448 : audit [DBG] from='client.15216 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:51.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:51 vm09 bash[21220]: audit 2026-03-07T07:02:49.744139+0000 mgr.vm07.yrfcuj (mgr.14201) 448 : audit [DBG] from='client.15216 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:51.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:51 vm09 bash[21220]: audit 2026-03-07T07:02:49.911673+0000 mgr.vm07.yrfcuj (mgr.14201) 449 : audit [DBG] from='client.15220 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:51.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:51 vm09 bash[21220]: audit 2026-03-07T07:02:49.911673+0000 mgr.vm07.yrfcuj (mgr.14201) 449 : audit [DBG] from='client.15220 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:51.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:51 vm09 bash[21220]: cluster 2026-03-07T07:02:50.496536+0000 mgr.vm07.yrfcuj (mgr.14201) 450 : cluster [DBG] pgmap v276: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:51.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:51 vm09 bash[21220]: cluster 2026-03-07T07:02:50.496536+0000 mgr.vm07.yrfcuj (mgr.14201) 450 : cluster [DBG] pgmap v276: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:51.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:51 vm07 bash[17031]: audit 2026-03-07T07:02:49.744139+0000 mgr.vm07.yrfcuj (mgr.14201) 448 : audit [DBG] from='client.15216 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:51.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:51 vm07 bash[17031]: audit 2026-03-07T07:02:49.744139+0000 mgr.vm07.yrfcuj (mgr.14201) 448 : audit [DBG] from='client.15216 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:51.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:51 vm07 bash[17031]: audit 2026-03-07T07:02:49.911673+0000 mgr.vm07.yrfcuj (mgr.14201) 449 : audit [DBG] from='client.15220 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:51.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:51 vm07 bash[17031]: audit 2026-03-07T07:02:49.911673+0000 mgr.vm07.yrfcuj (mgr.14201) 449 : audit [DBG] from='client.15220 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:51.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:51 vm07 bash[17031]: cluster 2026-03-07T07:02:50.496536+0000 mgr.vm07.yrfcuj (mgr.14201) 450 : cluster [DBG] pgmap v276: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:51.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:51 vm07 bash[17031]: cluster 2026-03-07T07:02:50.496536+0000 mgr.vm07.yrfcuj (mgr.14201) 450 : cluster [DBG] pgmap v276: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:52 vm09 bash[21220]: cluster 2026-03-07T07:02:52.497042+0000 mgr.vm07.yrfcuj (mgr.14201) 451 : cluster [DBG] pgmap v277: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:52 vm09 bash[21220]: cluster 2026-03-07T07:02:52.497042+0000 mgr.vm07.yrfcuj (mgr.14201) 451 : cluster [DBG] pgmap v277: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:52.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:52 vm07 bash[17031]: cluster 2026-03-07T07:02:52.497042+0000 mgr.vm07.yrfcuj (mgr.14201) 451 : cluster [DBG] pgmap v277: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:52.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:52 vm07 bash[17031]: cluster 2026-03-07T07:02:52.497042+0000 mgr.vm07.yrfcuj (mgr.14201) 451 : cluster [DBG] pgmap v277: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:54 vm09 bash[21220]: cluster 2026-03-07T07:02:54.497479+0000 mgr.vm07.yrfcuj (mgr.14201) 452 : cluster [DBG] pgmap v278: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:54 vm09 bash[21220]: cluster 2026-03-07T07:02:54.497479+0000 mgr.vm07.yrfcuj (mgr.14201) 452 : cluster [DBG] pgmap v278: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:54.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:54 vm07 bash[17031]: cluster 2026-03-07T07:02:54.497479+0000 mgr.vm07.yrfcuj (mgr.14201) 452 : cluster [DBG] pgmap v278: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:54.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:54 vm07 bash[17031]: cluster 2026-03-07T07:02:54.497479+0000 mgr.vm07.yrfcuj (mgr.14201) 452 : cluster [DBG] pgmap v278: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:55.324 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:02:55.478 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:02:55.479 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 53s ago 5m - - 2026-03-07T08:02:55.479 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 53s ago 5m 109M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:02:55.479 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (5m) 53s ago 5m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:02:55.479 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (5m) 53s ago 5m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:02:55.668 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:02:55.668 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:02:55.668 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:02:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:55 vm09 bash[21220]: audit 2026-03-07T07:02:55.132420+0000 mon.vm07 (mon.0) 913 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:02:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:55 vm09 bash[21220]: audit 2026-03-07T07:02:55.132420+0000 mon.vm07 (mon.0) 913 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:02:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:55 vm09 bash[21220]: audit 2026-03-07T07:02:55.293795+0000 mgr.vm07.yrfcuj (mgr.14201) 453 : audit [DBG] from='client.15228 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:55 vm09 bash[21220]: audit 2026-03-07T07:02:55.293795+0000 mgr.vm07.yrfcuj (mgr.14201) 453 : audit [DBG] from='client.15228 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:55 vm09 bash[21220]: audit 2026-03-07T07:02:55.475854+0000 mgr.vm07.yrfcuj (mgr.14201) 454 : audit [DBG] from='client.15232 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:55 vm09 bash[21220]: audit 2026-03-07T07:02:55.475854+0000 mgr.vm07.yrfcuj (mgr.14201) 454 : audit [DBG] from='client.15232 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:55.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:55 vm07 bash[17031]: audit 2026-03-07T07:02:55.132420+0000 mon.vm07 (mon.0) 913 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:02:55.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:55 vm07 bash[17031]: audit 2026-03-07T07:02:55.132420+0000 mon.vm07 (mon.0) 913 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:02:55.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:55 vm07 bash[17031]: audit 2026-03-07T07:02:55.293795+0000 mgr.vm07.yrfcuj (mgr.14201) 453 : audit [DBG] from='client.15228 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:55.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:55 vm07 bash[17031]: audit 2026-03-07T07:02:55.293795+0000 mgr.vm07.yrfcuj (mgr.14201) 453 : audit [DBG] from='client.15228 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:55.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:55 vm07 bash[17031]: audit 2026-03-07T07:02:55.475854+0000 mgr.vm07.yrfcuj (mgr.14201) 454 : audit [DBG] from='client.15232 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:55.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:55 vm07 bash[17031]: audit 2026-03-07T07:02:55.475854+0000 mgr.vm07.yrfcuj (mgr.14201) 454 : audit [DBG] from='client.15232 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:02:56.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:56 vm09 bash[21220]: audit 2026-03-07T07:02:55.667638+0000 mon.vm07 (mon.0) 914 : audit [DBG] from='client.? 192.168.123.107:0/853840829' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:56.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:56 vm09 bash[21220]: audit 2026-03-07T07:02:55.667638+0000 mon.vm07 (mon.0) 914 : audit [DBG] from='client.? 192.168.123.107:0/853840829' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:56.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:56 vm09 bash[21220]: cluster 2026-03-07T07:02:56.497843+0000 mgr.vm07.yrfcuj (mgr.14201) 455 : cluster [DBG] pgmap v279: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:56.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:56 vm09 bash[21220]: cluster 2026-03-07T07:02:56.497843+0000 mgr.vm07.yrfcuj (mgr.14201) 455 : cluster [DBG] pgmap v279: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:56.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:56 vm07 bash[17031]: audit 2026-03-07T07:02:55.667638+0000 mon.vm07 (mon.0) 914 : audit [DBG] from='client.? 192.168.123.107:0/853840829' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:56.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:56 vm07 bash[17031]: audit 2026-03-07T07:02:55.667638+0000 mon.vm07 (mon.0) 914 : audit [DBG] from='client.? 192.168.123.107:0/853840829' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:02:56.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:56 vm07 bash[17031]: cluster 2026-03-07T07:02:56.497843+0000 mgr.vm07.yrfcuj (mgr.14201) 455 : cluster [DBG] pgmap v279: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:56.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:56 vm07 bash[17031]: cluster 2026-03-07T07:02:56.497843+0000 mgr.vm07.yrfcuj (mgr.14201) 455 : cluster [DBG] pgmap v279: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:58.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:58 vm09 bash[21220]: cluster 2026-03-07T07:02:58.498301+0000 mgr.vm07.yrfcuj (mgr.14201) 456 : cluster [DBG] pgmap v280: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:58.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:02:58 vm09 bash[21220]: cluster 2026-03-07T07:02:58.498301+0000 mgr.vm07.yrfcuj (mgr.14201) 456 : cluster [DBG] pgmap v280: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:58.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:58 vm07 bash[17031]: cluster 2026-03-07T07:02:58.498301+0000 mgr.vm07.yrfcuj (mgr.14201) 456 : cluster [DBG] pgmap v280: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:02:58.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:02:58 vm07 bash[17031]: cluster 2026-03-07T07:02:58.498301+0000 mgr.vm07.yrfcuj (mgr.14201) 456 : cluster [DBG] pgmap v280: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:03:00.845 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:03:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:00 vm09 bash[21220]: cluster 2026-03-07T07:03:00.498685+0000 mgr.vm07.yrfcuj (mgr.14201) 457 : cluster [DBG] pgmap v281: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:00 vm09 bash[21220]: cluster 2026-03-07T07:03:00.498685+0000 mgr.vm07.yrfcuj (mgr.14201) 457 : cluster [DBG] pgmap v281: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:00.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:00 vm07 bash[17031]: cluster 2026-03-07T07:03:00.498685+0000 mgr.vm07.yrfcuj (mgr.14201) 457 : cluster [DBG] pgmap v281: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:00.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:00 vm07 bash[17031]: cluster 2026-03-07T07:03:00.498685+0000 mgr.vm07.yrfcuj (mgr.14201) 457 : cluster [DBG] pgmap v281: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:00.994 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:03:00.994 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 58s ago 5m - - 2026-03-07T08:03:00.994 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 58s ago 5m 109M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:03:00.994 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (5m) 59s ago 5m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:03:00.994 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (5m) 59s ago 5m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:03:01.187 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:03:01.187 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:03:01.187 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:03:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:01 vm09 bash[21220]: audit 2026-03-07T07:03:00.830153+0000 mgr.vm07.yrfcuj (mgr.14201) 458 : audit [DBG] from='client.15240 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:01 vm09 bash[21220]: audit 2026-03-07T07:03:00.830153+0000 mgr.vm07.yrfcuj (mgr.14201) 458 : audit [DBG] from='client.15240 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:01 vm09 bash[21220]: audit 2026-03-07T07:03:00.990735+0000 mgr.vm07.yrfcuj (mgr.14201) 459 : audit [DBG] from='client.15244 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:01 vm09 bash[21220]: audit 2026-03-07T07:03:00.990735+0000 mgr.vm07.yrfcuj (mgr.14201) 459 : audit [DBG] from='client.15244 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:01 vm09 bash[21220]: audit 2026-03-07T07:03:01.186425+0000 mon.vm07 (mon.0) 915 : audit [DBG] from='client.? 192.168.123.107:0/587370624' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:01 vm09 bash[21220]: audit 2026-03-07T07:03:01.186425+0000 mon.vm07 (mon.0) 915 : audit [DBG] from='client.? 192.168.123.107:0/587370624' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:01.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:01 vm07 bash[17031]: audit 2026-03-07T07:03:00.830153+0000 mgr.vm07.yrfcuj (mgr.14201) 458 : audit [DBG] from='client.15240 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:01.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:01 vm07 bash[17031]: audit 2026-03-07T07:03:00.830153+0000 mgr.vm07.yrfcuj (mgr.14201) 458 : audit [DBG] from='client.15240 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:01.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:01 vm07 bash[17031]: audit 2026-03-07T07:03:00.990735+0000 mgr.vm07.yrfcuj (mgr.14201) 459 : audit [DBG] from='client.15244 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:01.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:01 vm07 bash[17031]: audit 2026-03-07T07:03:00.990735+0000 mgr.vm07.yrfcuj (mgr.14201) 459 : audit [DBG] from='client.15244 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:01.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:01 vm07 bash[17031]: audit 2026-03-07T07:03:01.186425+0000 mon.vm07 (mon.0) 915 : audit [DBG] from='client.? 192.168.123.107:0/587370624' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:01.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:01 vm07 bash[17031]: audit 2026-03-07T07:03:01.186425+0000 mon.vm07 (mon.0) 915 : audit [DBG] from='client.? 192.168.123.107:0/587370624' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:02.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:02 vm07 bash[17031]: cluster 2026-03-07T07:03:02.499116+0000 mgr.vm07.yrfcuj (mgr.14201) 460 : cluster [DBG] pgmap v282: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:02.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:02 vm07 bash[17031]: cluster 2026-03-07T07:03:02.499116+0000 mgr.vm07.yrfcuj (mgr.14201) 460 : cluster [DBG] pgmap v282: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:02.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:02 vm07 bash[17031]: audit 2026-03-07T07:03:02.531297+0000 mon.vm07 (mon.0) 916 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:02.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:02 vm07 bash[17031]: audit 2026-03-07T07:03:02.531297+0000 mon.vm07 (mon.0) 916 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:03.085 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:02 vm09 bash[21220]: cluster 2026-03-07T07:03:02.499116+0000 mgr.vm07.yrfcuj (mgr.14201) 460 : cluster [DBG] pgmap v282: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:03.085 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:02 vm09 bash[21220]: cluster 2026-03-07T07:03:02.499116+0000 mgr.vm07.yrfcuj (mgr.14201) 460 : cluster [DBG] pgmap v282: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:03.085 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:02 vm09 bash[21220]: audit 2026-03-07T07:03:02.531297+0000 mon.vm07 (mon.0) 916 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:03.085 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:02 vm09 bash[21220]: audit 2026-03-07T07:03:02.531297+0000 mon.vm07 (mon.0) 916 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:04.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:04 vm09 bash[21220]: cluster 2026-03-07T07:03:04.499553+0000 mgr.vm07.yrfcuj (mgr.14201) 461 : cluster [DBG] pgmap v283: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:04.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:04 vm09 bash[21220]: cluster 2026-03-07T07:03:04.499553+0000 mgr.vm07.yrfcuj (mgr.14201) 461 : cluster [DBG] pgmap v283: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:04.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:04 vm07 bash[17031]: cluster 2026-03-07T07:03:04.499553+0000 mgr.vm07.yrfcuj (mgr.14201) 461 : cluster [DBG] pgmap v283: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:04.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:04 vm07 bash[17031]: cluster 2026-03-07T07:03:04.499553+0000 mgr.vm07.yrfcuj (mgr.14201) 461 : cluster [DBG] pgmap v283: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:06.373 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:03:06.522 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:03:06.522 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 64s ago 5m - - 2026-03-07T08:03:06.522 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 64s ago 5m 109M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:03:06.522 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (5m) 64s ago 5m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:03:06.522 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (5m) 64s ago 5m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:03:06.714 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:03:06.714 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:03:06.714 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:03:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:06 vm09 bash[21220]: audit 2026-03-07T07:03:06.357550+0000 mgr.vm07.yrfcuj (mgr.14201) 462 : audit [DBG] from='client.15252 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:06 vm09 bash[21220]: audit 2026-03-07T07:03:06.357550+0000 mgr.vm07.yrfcuj (mgr.14201) 462 : audit [DBG] from='client.15252 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:06 vm09 bash[21220]: cluster 2026-03-07T07:03:06.499910+0000 mgr.vm07.yrfcuj (mgr.14201) 463 : cluster [DBG] pgmap v284: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:06 vm09 bash[21220]: cluster 2026-03-07T07:03:06.499910+0000 mgr.vm07.yrfcuj (mgr.14201) 463 : cluster [DBG] pgmap v284: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:06 vm09 bash[21220]: audit 2026-03-07T07:03:06.519657+0000 mgr.vm07.yrfcuj (mgr.14201) 464 : audit [DBG] from='client.15256 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:06 vm09 bash[21220]: audit 2026-03-07T07:03:06.519657+0000 mgr.vm07.yrfcuj (mgr.14201) 464 : audit [DBG] from='client.15256 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:06.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:06 vm07 bash[17031]: audit 2026-03-07T07:03:06.357550+0000 mgr.vm07.yrfcuj (mgr.14201) 462 : audit [DBG] from='client.15252 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:06.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:06 vm07 bash[17031]: audit 2026-03-07T07:03:06.357550+0000 mgr.vm07.yrfcuj (mgr.14201) 462 : audit [DBG] from='client.15252 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:06.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:06 vm07 bash[17031]: cluster 2026-03-07T07:03:06.499910+0000 mgr.vm07.yrfcuj (mgr.14201) 463 : cluster [DBG] pgmap v284: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:06.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:06 vm07 bash[17031]: cluster 2026-03-07T07:03:06.499910+0000 mgr.vm07.yrfcuj (mgr.14201) 463 : cluster [DBG] pgmap v284: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:06.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:06 vm07 bash[17031]: audit 2026-03-07T07:03:06.519657+0000 mgr.vm07.yrfcuj (mgr.14201) 464 : audit [DBG] from='client.15256 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:06.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:06 vm07 bash[17031]: audit 2026-03-07T07:03:06.519657+0000 mgr.vm07.yrfcuj (mgr.14201) 464 : audit [DBG] from='client.15256 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:07.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:07 vm09 bash[21220]: audit 2026-03-07T07:03:06.713677+0000 mon.vm07 (mon.0) 917 : audit [DBG] from='client.? 192.168.123.107:0/4107004368' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:07.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:07 vm09 bash[21220]: audit 2026-03-07T07:03:06.713677+0000 mon.vm07 (mon.0) 917 : audit [DBG] from='client.? 192.168.123.107:0/4107004368' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:07.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:07 vm07 bash[17031]: audit 2026-03-07T07:03:06.713677+0000 mon.vm07 (mon.0) 917 : audit [DBG] from='client.? 192.168.123.107:0/4107004368' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:07.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:07 vm07 bash[17031]: audit 2026-03-07T07:03:06.713677+0000 mon.vm07 (mon.0) 917 : audit [DBG] from='client.? 192.168.123.107:0/4107004368' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:09.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:09 vm09 bash[21220]: audit 2026-03-07T07:03:08.042537+0000 mon.vm07 (mon.0) 918 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:09.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:09 vm09 bash[21220]: audit 2026-03-07T07:03:08.042537+0000 mon.vm07 (mon.0) 918 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:09.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:09 vm09 bash[21220]: audit 2026-03-07T07:03:08.048190+0000 mon.vm07 (mon.0) 919 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:09.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:09 vm09 bash[21220]: audit 2026-03-07T07:03:08.048190+0000 mon.vm07 (mon.0) 919 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:09.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:09 vm09 bash[21220]: audit 2026-03-07T07:03:08.342430+0000 mon.vm07 (mon.0) 920 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:09.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:09 vm09 bash[21220]: audit 2026-03-07T07:03:08.342430+0000 mon.vm07 (mon.0) 920 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:09.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:09 vm09 bash[21220]: audit 2026-03-07T07:03:08.343051+0000 mon.vm07 (mon.0) 921 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:09.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:09 vm09 bash[21220]: audit 2026-03-07T07:03:08.343051+0000 mon.vm07 (mon.0) 921 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:09.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:09 vm09 bash[21220]: cluster 2026-03-07T07:03:08.344085+0000 mgr.vm07.yrfcuj (mgr.14201) 465 : cluster [DBG] pgmap v285: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:09.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:09 vm09 bash[21220]: cluster 2026-03-07T07:03:08.344085+0000 mgr.vm07.yrfcuj (mgr.14201) 465 : cluster [DBG] pgmap v285: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:09.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:09 vm09 bash[21220]: cluster 2026-03-07T07:03:08.344231+0000 mgr.vm07.yrfcuj (mgr.14201) 466 : cluster [DBG] pgmap v286: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:09.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:09 vm09 bash[21220]: cluster 2026-03-07T07:03:08.344231+0000 mgr.vm07.yrfcuj (mgr.14201) 466 : cluster [DBG] pgmap v286: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:09.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:09 vm09 bash[21220]: audit 2026-03-07T07:03:08.347992+0000 mon.vm07 (mon.0) 922 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:09.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:09 vm09 bash[21220]: audit 2026-03-07T07:03:08.347992+0000 mon.vm07 (mon.0) 922 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:09.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:09 vm09 bash[21220]: audit 2026-03-07T07:03:08.349279+0000 mon.vm07 (mon.0) 923 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:09.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:09 vm09 bash[21220]: audit 2026-03-07T07:03:08.349279+0000 mon.vm07 (mon.0) 923 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:09.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:09 vm07 bash[17031]: audit 2026-03-07T07:03:08.042537+0000 mon.vm07 (mon.0) 918 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:09.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:09 vm07 bash[17031]: audit 2026-03-07T07:03:08.042537+0000 mon.vm07 (mon.0) 918 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:09.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:09 vm07 bash[17031]: audit 2026-03-07T07:03:08.048190+0000 mon.vm07 (mon.0) 919 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:09.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:09 vm07 bash[17031]: audit 2026-03-07T07:03:08.048190+0000 mon.vm07 (mon.0) 919 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:09.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:09 vm07 bash[17031]: audit 2026-03-07T07:03:08.342430+0000 mon.vm07 (mon.0) 920 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:09.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:09 vm07 bash[17031]: audit 2026-03-07T07:03:08.342430+0000 mon.vm07 (mon.0) 920 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:09.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:09 vm07 bash[17031]: audit 2026-03-07T07:03:08.343051+0000 mon.vm07 (mon.0) 921 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:09.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:09 vm07 bash[17031]: audit 2026-03-07T07:03:08.343051+0000 mon.vm07 (mon.0) 921 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:09.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:09 vm07 bash[17031]: cluster 2026-03-07T07:03:08.344085+0000 mgr.vm07.yrfcuj (mgr.14201) 465 : cluster [DBG] pgmap v285: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:09.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:09 vm07 bash[17031]: cluster 2026-03-07T07:03:08.344085+0000 mgr.vm07.yrfcuj (mgr.14201) 465 : cluster [DBG] pgmap v285: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:09.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:09 vm07 bash[17031]: cluster 2026-03-07T07:03:08.344231+0000 mgr.vm07.yrfcuj (mgr.14201) 466 : cluster [DBG] pgmap v286: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:09.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:09 vm07 bash[17031]: cluster 2026-03-07T07:03:08.344231+0000 mgr.vm07.yrfcuj (mgr.14201) 466 : cluster [DBG] pgmap v286: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:09.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:09 vm07 bash[17031]: audit 2026-03-07T07:03:08.347992+0000 mon.vm07 (mon.0) 922 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:09.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:09 vm07 bash[17031]: audit 2026-03-07T07:03:08.347992+0000 mon.vm07 (mon.0) 922 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:09.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:09 vm07 bash[17031]: audit 2026-03-07T07:03:08.349279+0000 mon.vm07 (mon.0) 923 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:09.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:09 vm07 bash[17031]: audit 2026-03-07T07:03:08.349279+0000 mon.vm07 (mon.0) 923 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:10.464 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:10 vm07 bash[17031]: audit 2026-03-07T07:03:10.132615+0000 mon.vm07 (mon.0) 924 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:03:10.464 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:10 vm07 bash[17031]: audit 2026-03-07T07:03:10.132615+0000 mon.vm07 (mon.0) 924 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:03:10.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:10 vm09 bash[21220]: audit 2026-03-07T07:03:10.132615+0000 mon.vm07 (mon.0) 924 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:03:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:10 vm09 bash[21220]: audit 2026-03-07T07:03:10.132615+0000 mon.vm07 (mon.0) 924 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:03:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:11 vm09 bash[21220]: cluster 2026-03-07T07:03:10.344698+0000 mgr.vm07.yrfcuj (mgr.14201) 467 : cluster [DBG] pgmap v287: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:11 vm09 bash[21220]: cluster 2026-03-07T07:03:10.344698+0000 mgr.vm07.yrfcuj (mgr.14201) 467 : cluster [DBG] pgmap v287: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:11.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:11 vm07 bash[17031]: cluster 2026-03-07T07:03:10.344698+0000 mgr.vm07.yrfcuj (mgr.14201) 467 : cluster [DBG] pgmap v287: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:11.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:11 vm07 bash[17031]: cluster 2026-03-07T07:03:10.344698+0000 mgr.vm07.yrfcuj (mgr.14201) 467 : cluster [DBG] pgmap v287: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:11.897 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:03:12.049 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:03:12.050 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 4s ago 5m - - 2026-03-07T08:03:12.050 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 4s ago 5m 114M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:03:12.050 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (5m) 70s ago 5m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:03:12.050 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (5m) 70s ago 5m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:03:12.231 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:03:12.231 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:03:12.231 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:03:12.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:12 vm09 bash[21220]: audit 2026-03-07T07:03:12.230853+0000 mon.vm07 (mon.0) 925 : audit [DBG] from='client.? 192.168.123.107:0/3550163162' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:12.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:12 vm09 bash[21220]: audit 2026-03-07T07:03:12.230853+0000 mon.vm07 (mon.0) 925 : audit [DBG] from='client.? 192.168.123.107:0/3550163162' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:12.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:12 vm07 bash[17031]: audit 2026-03-07T07:03:12.230853+0000 mon.vm07 (mon.0) 925 : audit [DBG] from='client.? 192.168.123.107:0/3550163162' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:12.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:12 vm07 bash[17031]: audit 2026-03-07T07:03:12.230853+0000 mon.vm07 (mon.0) 925 : audit [DBG] from='client.? 192.168.123.107:0/3550163162' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:13 vm09 bash[21220]: audit 2026-03-07T07:03:11.882099+0000 mgr.vm07.yrfcuj (mgr.14201) 468 : audit [DBG] from='client.15264 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:13 vm09 bash[21220]: audit 2026-03-07T07:03:11.882099+0000 mgr.vm07.yrfcuj (mgr.14201) 468 : audit [DBG] from='client.15264 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:13 vm09 bash[21220]: audit 2026-03-07T07:03:12.047125+0000 mgr.vm07.yrfcuj (mgr.14201) 469 : audit [DBG] from='client.15268 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:13 vm09 bash[21220]: audit 2026-03-07T07:03:12.047125+0000 mgr.vm07.yrfcuj (mgr.14201) 469 : audit [DBG] from='client.15268 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:13 vm09 bash[21220]: cluster 2026-03-07T07:03:12.345162+0000 mgr.vm07.yrfcuj (mgr.14201) 470 : cluster [DBG] pgmap v288: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:13 vm09 bash[21220]: cluster 2026-03-07T07:03:12.345162+0000 mgr.vm07.yrfcuj (mgr.14201) 470 : cluster [DBG] pgmap v288: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:13.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:13 vm07 bash[17031]: audit 2026-03-07T07:03:11.882099+0000 mgr.vm07.yrfcuj (mgr.14201) 468 : audit [DBG] from='client.15264 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:13.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:13 vm07 bash[17031]: audit 2026-03-07T07:03:11.882099+0000 mgr.vm07.yrfcuj (mgr.14201) 468 : audit [DBG] from='client.15264 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:13.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:13 vm07 bash[17031]: audit 2026-03-07T07:03:12.047125+0000 mgr.vm07.yrfcuj (mgr.14201) 469 : audit [DBG] from='client.15268 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:13.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:13 vm07 bash[17031]: audit 2026-03-07T07:03:12.047125+0000 mgr.vm07.yrfcuj (mgr.14201) 469 : audit [DBG] from='client.15268 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:13.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:13 vm07 bash[17031]: cluster 2026-03-07T07:03:12.345162+0000 mgr.vm07.yrfcuj (mgr.14201) 470 : cluster [DBG] pgmap v288: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:13.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:13 vm07 bash[17031]: cluster 2026-03-07T07:03:12.345162+0000 mgr.vm07.yrfcuj (mgr.14201) 470 : cluster [DBG] pgmap v288: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:15.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:15 vm07 bash[17031]: cluster 2026-03-07T07:03:14.345589+0000 mgr.vm07.yrfcuj (mgr.14201) 471 : cluster [DBG] pgmap v289: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:15.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:15 vm07 bash[17031]: cluster 2026-03-07T07:03:14.345589+0000 mgr.vm07.yrfcuj (mgr.14201) 471 : cluster [DBG] pgmap v289: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:15.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:15 vm09 bash[21220]: cluster 2026-03-07T07:03:14.345589+0000 mgr.vm07.yrfcuj (mgr.14201) 471 : cluster [DBG] pgmap v289: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:15.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:15 vm09 bash[21220]: cluster 2026-03-07T07:03:14.345589+0000 mgr.vm07.yrfcuj (mgr.14201) 471 : cluster [DBG] pgmap v289: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:17.405 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to stop 2026-03-07T08:03:17.561 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:03:17.561 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 9s ago 5m - - 2026-03-07T08:03:17.561 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 9s ago 5m 114M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:03:17.561 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (5m) 75s ago 5m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:03:17.561 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (5m) 75s ago 5m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:03:17.761 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:03:17.762 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:03:17.762 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:03:17.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:17 vm09 bash[21220]: cluster 2026-03-07T07:03:16.345976+0000 mgr.vm07.yrfcuj (mgr.14201) 472 : cluster [DBG] pgmap v290: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:17.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:17 vm09 bash[21220]: cluster 2026-03-07T07:03:16.345976+0000 mgr.vm07.yrfcuj (mgr.14201) 472 : cluster [DBG] pgmap v290: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:17.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:17 vm07 bash[17031]: cluster 2026-03-07T07:03:16.345976+0000 mgr.vm07.yrfcuj (mgr.14201) 472 : cluster [DBG] pgmap v290: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:17.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:17 vm07 bash[17031]: cluster 2026-03-07T07:03:16.345976+0000 mgr.vm07.yrfcuj (mgr.14201) 472 : cluster [DBG] pgmap v290: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:03:18.587 INFO:teuthology.orchestra.run.vm07.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-07T08:03:18.587 INFO:teuthology.orchestra.run.vm07.stderr: Dload Upload Total Spent Left Speed 2026-03-07T08:03:18.588 INFO:teuthology.orchestra.run.vm07.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 93500 0 --:--:-- --:--:-- --:--:-- 93500 2026-03-07T08:03:18.758 INFO:teuthology.orchestra.run.vm07.stdout:anonymousScheduled to start rgw.foo.vm07.bgfaka on host 'vm07' 2026-03-07T08:03:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:18 vm09 bash[21220]: audit 2026-03-07T07:03:17.391344+0000 mgr.vm07.yrfcuj (mgr.14201) 473 : audit [DBG] from='client.15276 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:18 vm09 bash[21220]: audit 2026-03-07T07:03:17.391344+0000 mgr.vm07.yrfcuj (mgr.14201) 473 : audit [DBG] from='client.15276 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:18 vm09 bash[21220]: audit 2026-03-07T07:03:17.761401+0000 mon.vm07 (mon.0) 926 : audit [DBG] from='client.? 192.168.123.107:0/590719799' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:18 vm09 bash[21220]: audit 2026-03-07T07:03:17.761401+0000 mon.vm07 (mon.0) 926 : audit [DBG] from='client.? 192.168.123.107:0/590719799' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:18.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:18 vm07 bash[17031]: audit 2026-03-07T07:03:17.391344+0000 mgr.vm07.yrfcuj (mgr.14201) 473 : audit [DBG] from='client.15276 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:18.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:18 vm07 bash[17031]: audit 2026-03-07T07:03:17.391344+0000 mgr.vm07.yrfcuj (mgr.14201) 473 : audit [DBG] from='client.15276 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:18.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:18 vm07 bash[17031]: audit 2026-03-07T07:03:17.761401+0000 mon.vm07 (mon.0) 926 : audit [DBG] from='client.? 192.168.123.107:0/590719799' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:18.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:18 vm07 bash[17031]: audit 2026-03-07T07:03:17.761401+0000 mon.vm07 (mon.0) 926 : audit [DBG] from='client.? 192.168.123.107:0/590719799' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:18.949 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to start 2026-03-07T08:03:19.138 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:03:19.138 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 11s ago 5m - - 2026-03-07T08:03:19.138 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 11s ago 5m 114M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:03:19.138 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (5m) 77s ago 5m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:03:19.138 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (5m) 77s ago 5m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:03:19.335 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:03:19.335 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:03:19.335 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: audit 2026-03-07T07:03:17.558385+0000 mgr.vm07.yrfcuj (mgr.14201) 474 : audit [DBG] from='client.15280 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: audit 2026-03-07T07:03:17.558385+0000 mgr.vm07.yrfcuj (mgr.14201) 474 : audit [DBG] from='client.15280 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: cluster 2026-03-07T07:03:18.346445+0000 mgr.vm07.yrfcuj (mgr.14201) 475 : cluster [DBG] pgmap v291: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 307 B/s rd, 614 B/s wr, 0 op/s 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: cluster 2026-03-07T07:03:18.346445+0000 mgr.vm07.yrfcuj (mgr.14201) 475 : cluster [DBG] pgmap v291: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 307 B/s rd, 614 B/s wr, 0 op/s 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: audit 2026-03-07T07:03:18.751318+0000 mon.vm07 (mon.0) 927 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: audit 2026-03-07T07:03:18.751318+0000 mon.vm07 (mon.0) 927 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: audit 2026-03-07T07:03:18.756995+0000 mon.vm07 (mon.0) 928 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: audit 2026-03-07T07:03:18.756995+0000 mon.vm07 (mon.0) 928 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: audit 2026-03-07T07:03:18.757977+0000 mon.vm07 (mon.0) 929 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: audit 2026-03-07T07:03:18.757977+0000 mon.vm07 (mon.0) 929 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: audit 2026-03-07T07:03:18.759145+0000 mon.vm07 (mon.0) 930 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: audit 2026-03-07T07:03:18.759145+0000 mon.vm07 (mon.0) 930 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: audit 2026-03-07T07:03:18.759558+0000 mon.vm07 (mon.0) 931 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: audit 2026-03-07T07:03:18.759558+0000 mon.vm07 (mon.0) 931 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: audit 2026-03-07T07:03:18.764515+0000 mon.vm07 (mon.0) 932 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: audit 2026-03-07T07:03:18.764515+0000 mon.vm07 (mon.0) 932 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: audit 2026-03-07T07:03:18.765995+0000 mon.vm07 (mon.0) 933 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: audit 2026-03-07T07:03:18.765995+0000 mon.vm07 (mon.0) 933 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: audit 2026-03-07T07:03:19.335095+0000 mon.vm07 (mon.0) 934 : audit [DBG] from='client.? 192.168.123.107:0/1376639252' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:19 vm09 bash[21220]: audit 2026-03-07T07:03:19.335095+0000 mon.vm07 (mon.0) 934 : audit [DBG] from='client.? 192.168.123.107:0/1376639252' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: audit 2026-03-07T07:03:17.558385+0000 mgr.vm07.yrfcuj (mgr.14201) 474 : audit [DBG] from='client.15280 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: audit 2026-03-07T07:03:17.558385+0000 mgr.vm07.yrfcuj (mgr.14201) 474 : audit [DBG] from='client.15280 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: cluster 2026-03-07T07:03:18.346445+0000 mgr.vm07.yrfcuj (mgr.14201) 475 : cluster [DBG] pgmap v291: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 307 B/s rd, 614 B/s wr, 0 op/s 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: cluster 2026-03-07T07:03:18.346445+0000 mgr.vm07.yrfcuj (mgr.14201) 475 : cluster [DBG] pgmap v291: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 307 B/s rd, 614 B/s wr, 0 op/s 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: audit 2026-03-07T07:03:18.751318+0000 mon.vm07 (mon.0) 927 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: audit 2026-03-07T07:03:18.751318+0000 mon.vm07 (mon.0) 927 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: audit 2026-03-07T07:03:18.756995+0000 mon.vm07 (mon.0) 928 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: audit 2026-03-07T07:03:18.756995+0000 mon.vm07 (mon.0) 928 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: audit 2026-03-07T07:03:18.757977+0000 mon.vm07 (mon.0) 929 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: audit 2026-03-07T07:03:18.757977+0000 mon.vm07 (mon.0) 929 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: audit 2026-03-07T07:03:18.759145+0000 mon.vm07 (mon.0) 930 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: audit 2026-03-07T07:03:18.759145+0000 mon.vm07 (mon.0) 930 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: audit 2026-03-07T07:03:18.759558+0000 mon.vm07 (mon.0) 931 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: audit 2026-03-07T07:03:18.759558+0000 mon.vm07 (mon.0) 931 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: audit 2026-03-07T07:03:18.764515+0000 mon.vm07 (mon.0) 932 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: audit 2026-03-07T07:03:18.764515+0000 mon.vm07 (mon.0) 932 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: audit 2026-03-07T07:03:18.765995+0000 mon.vm07 (mon.0) 933 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: audit 2026-03-07T07:03:18.765995+0000 mon.vm07 (mon.0) 933 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: audit 2026-03-07T07:03:19.335095+0000 mon.vm07 (mon.0) 934 : audit [DBG] from='client.? 192.168.123.107:0/1376639252' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:19.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:19 vm07 bash[17031]: audit 2026-03-07T07:03:19.335095+0000 mon.vm07 (mon.0) 934 : audit [DBG] from='client.? 192.168.123.107:0/1376639252' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:20 vm09 bash[21220]: audit 2026-03-07T07:03:18.745521+0000 mgr.vm07.yrfcuj (mgr.14201) 476 : audit [DBG] from='client.24825 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "rgw.foo.vm07.bgfaka", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:20 vm09 bash[21220]: audit 2026-03-07T07:03:18.745521+0000 mgr.vm07.yrfcuj (mgr.14201) 476 : audit [DBG] from='client.24825 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "rgw.foo.vm07.bgfaka", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:20 vm09 bash[21220]: cephadm 2026-03-07T07:03:18.745871+0000 mgr.vm07.yrfcuj (mgr.14201) 477 : cephadm [INF] Schedule start daemon rgw.foo.vm07.bgfaka 2026-03-07T08:03:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:20 vm09 bash[21220]: cephadm 2026-03-07T07:03:18.745871+0000 mgr.vm07.yrfcuj (mgr.14201) 477 : cephadm [INF] Schedule start daemon rgw.foo.vm07.bgfaka 2026-03-07T08:03:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:20 vm09 bash[21220]: cluster 2026-03-07T07:03:18.760584+0000 mgr.vm07.yrfcuj (mgr.14201) 478 : cluster [DBG] pgmap v292: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:20 vm09 bash[21220]: cluster 2026-03-07T07:03:18.760584+0000 mgr.vm07.yrfcuj (mgr.14201) 478 : cluster [DBG] pgmap v292: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:20 vm09 bash[21220]: audit 2026-03-07T07:03:18.929548+0000 mgr.vm07.yrfcuj (mgr.14201) 479 : audit [DBG] from='client.15292 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:20 vm09 bash[21220]: audit 2026-03-07T07:03:18.929548+0000 mgr.vm07.yrfcuj (mgr.14201) 479 : audit [DBG] from='client.15292 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:20 vm09 bash[21220]: audit 2026-03-07T07:03:19.135509+0000 mgr.vm07.yrfcuj (mgr.14201) 480 : audit [DBG] from='client.15296 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:20 vm09 bash[21220]: audit 2026-03-07T07:03:19.135509+0000 mgr.vm07.yrfcuj (mgr.14201) 480 : audit [DBG] from='client.15296 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:20.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:20 vm07 bash[17031]: audit 2026-03-07T07:03:18.745521+0000 mgr.vm07.yrfcuj (mgr.14201) 476 : audit [DBG] from='client.24825 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "rgw.foo.vm07.bgfaka", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:20.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:20 vm07 bash[17031]: audit 2026-03-07T07:03:18.745521+0000 mgr.vm07.yrfcuj (mgr.14201) 476 : audit [DBG] from='client.24825 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "rgw.foo.vm07.bgfaka", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:20.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:20 vm07 bash[17031]: cephadm 2026-03-07T07:03:18.745871+0000 mgr.vm07.yrfcuj (mgr.14201) 477 : cephadm [INF] Schedule start daemon rgw.foo.vm07.bgfaka 2026-03-07T08:03:20.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:20 vm07 bash[17031]: cephadm 2026-03-07T07:03:18.745871+0000 mgr.vm07.yrfcuj (mgr.14201) 477 : cephadm [INF] Schedule start daemon rgw.foo.vm07.bgfaka 2026-03-07T08:03:20.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:20 vm07 bash[17031]: cluster 2026-03-07T07:03:18.760584+0000 mgr.vm07.yrfcuj (mgr.14201) 478 : cluster [DBG] pgmap v292: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:20.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:20 vm07 bash[17031]: cluster 2026-03-07T07:03:18.760584+0000 mgr.vm07.yrfcuj (mgr.14201) 478 : cluster [DBG] pgmap v292: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:20.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:20 vm07 bash[17031]: audit 2026-03-07T07:03:18.929548+0000 mgr.vm07.yrfcuj (mgr.14201) 479 : audit [DBG] from='client.15292 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:20.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:20 vm07 bash[17031]: audit 2026-03-07T07:03:18.929548+0000 mgr.vm07.yrfcuj (mgr.14201) 479 : audit [DBG] from='client.15292 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:20.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:20 vm07 bash[17031]: audit 2026-03-07T07:03:19.135509+0000 mgr.vm07.yrfcuj (mgr.14201) 480 : audit [DBG] from='client.15296 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:20.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:20 vm07 bash[17031]: audit 2026-03-07T07:03:19.135509+0000 mgr.vm07.yrfcuj (mgr.14201) 480 : audit [DBG] from='client.15296 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:22.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:22 vm09 bash[21220]: cluster 2026-03-07T07:03:20.761040+0000 mgr.vm07.yrfcuj (mgr.14201) 481 : cluster [DBG] pgmap v293: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:22.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:22 vm09 bash[21220]: cluster 2026-03-07T07:03:20.761040+0000 mgr.vm07.yrfcuj (mgr.14201) 481 : cluster [DBG] pgmap v293: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:22.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:22 vm07 bash[17031]: cluster 2026-03-07T07:03:20.761040+0000 mgr.vm07.yrfcuj (mgr.14201) 481 : cluster [DBG] pgmap v293: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:22.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:22 vm07 bash[17031]: cluster 2026-03-07T07:03:20.761040+0000 mgr.vm07.yrfcuj (mgr.14201) 481 : cluster [DBG] pgmap v293: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:24.519 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to start 2026-03-07T08:03:24.670 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:03:24.671 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 16s ago 5m - - 2026-03-07T08:03:24.671 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 16s ago 5m 114M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:03:24.671 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (5m) 82s ago 5m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:03:24.671 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (5m) 82s ago 5m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:03:24.860 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:03:24.860 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:03:24.860 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:03:24.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:24 vm09 bash[21220]: cluster 2026-03-07T07:03:22.761632+0000 mgr.vm07.yrfcuj (mgr.14201) 482 : cluster [DBG] pgmap v294: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:24.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:24 vm09 bash[21220]: cluster 2026-03-07T07:03:22.761632+0000 mgr.vm07.yrfcuj (mgr.14201) 482 : cluster [DBG] pgmap v294: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:24.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:24 vm07 bash[17031]: cluster 2026-03-07T07:03:22.761632+0000 mgr.vm07.yrfcuj (mgr.14201) 482 : cluster [DBG] pgmap v294: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:24.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:24 vm07 bash[17031]: cluster 2026-03-07T07:03:22.761632+0000 mgr.vm07.yrfcuj (mgr.14201) 482 : cluster [DBG] pgmap v294: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:25 vm09 bash[21220]: audit 2026-03-07T07:03:24.503085+0000 mgr.vm07.yrfcuj (mgr.14201) 483 : audit [DBG] from='client.15304 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:25 vm09 bash[21220]: audit 2026-03-07T07:03:24.503085+0000 mgr.vm07.yrfcuj (mgr.14201) 483 : audit [DBG] from='client.15304 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:25 vm09 bash[21220]: audit 2026-03-07T07:03:24.860375+0000 mon.vm07 (mon.0) 935 : audit [DBG] from='client.? 192.168.123.107:0/2416736982' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:25 vm09 bash[21220]: audit 2026-03-07T07:03:24.860375+0000 mon.vm07 (mon.0) 935 : audit [DBG] from='client.? 192.168.123.107:0/2416736982' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:25 vm09 bash[21220]: audit 2026-03-07T07:03:25.134628+0000 mon.vm07 (mon.0) 936 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:03:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:25 vm09 bash[21220]: audit 2026-03-07T07:03:25.134628+0000 mon.vm07 (mon.0) 936 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:03:25.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:25 vm07 bash[17031]: audit 2026-03-07T07:03:24.503085+0000 mgr.vm07.yrfcuj (mgr.14201) 483 : audit [DBG] from='client.15304 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:25.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:25 vm07 bash[17031]: audit 2026-03-07T07:03:24.503085+0000 mgr.vm07.yrfcuj (mgr.14201) 483 : audit [DBG] from='client.15304 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:25.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:25 vm07 bash[17031]: audit 2026-03-07T07:03:24.860375+0000 mon.vm07 (mon.0) 935 : audit [DBG] from='client.? 192.168.123.107:0/2416736982' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:25.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:25 vm07 bash[17031]: audit 2026-03-07T07:03:24.860375+0000 mon.vm07 (mon.0) 935 : audit [DBG] from='client.? 192.168.123.107:0/2416736982' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:25.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:25 vm07 bash[17031]: audit 2026-03-07T07:03:25.134628+0000 mon.vm07 (mon.0) 936 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:03:25.903 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:25 vm07 bash[17031]: audit 2026-03-07T07:03:25.134628+0000 mon.vm07 (mon.0) 936 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:03:27.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:26 vm09 bash[21220]: audit 2026-03-07T07:03:24.668030+0000 mgr.vm07.yrfcuj (mgr.14201) 484 : audit [DBG] from='client.15308 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:27.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:26 vm09 bash[21220]: audit 2026-03-07T07:03:24.668030+0000 mgr.vm07.yrfcuj (mgr.14201) 484 : audit [DBG] from='client.15308 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:27.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:26 vm09 bash[21220]: cluster 2026-03-07T07:03:24.762070+0000 mgr.vm07.yrfcuj (mgr.14201) 485 : cluster [DBG] pgmap v295: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:27.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:26 vm09 bash[21220]: cluster 2026-03-07T07:03:24.762070+0000 mgr.vm07.yrfcuj (mgr.14201) 485 : cluster [DBG] pgmap v295: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:27.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:26 vm07 bash[17031]: audit 2026-03-07T07:03:24.668030+0000 mgr.vm07.yrfcuj (mgr.14201) 484 : audit [DBG] from='client.15308 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:27.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:26 vm07 bash[17031]: audit 2026-03-07T07:03:24.668030+0000 mgr.vm07.yrfcuj (mgr.14201) 484 : audit [DBG] from='client.15308 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:27.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:26 vm07 bash[17031]: cluster 2026-03-07T07:03:24.762070+0000 mgr.vm07.yrfcuj (mgr.14201) 485 : cluster [DBG] pgmap v295: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:27.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:26 vm07 bash[17031]: cluster 2026-03-07T07:03:24.762070+0000 mgr.vm07.yrfcuj (mgr.14201) 485 : cluster [DBG] pgmap v295: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:27.981 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:27 vm07 bash[17031]: cluster 2026-03-07T07:03:26.762507+0000 mgr.vm07.yrfcuj (mgr.14201) 486 : cluster [DBG] pgmap v296: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:27.981 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:27 vm07 bash[17031]: cluster 2026-03-07T07:03:26.762507+0000 mgr.vm07.yrfcuj (mgr.14201) 486 : cluster [DBG] pgmap v296: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:28.085 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:27 vm09 bash[21220]: cluster 2026-03-07T07:03:26.762507+0000 mgr.vm07.yrfcuj (mgr.14201) 486 : cluster [DBG] pgmap v296: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:28.085 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:27 vm09 bash[21220]: cluster 2026-03-07T07:03:26.762507+0000 mgr.vm07.yrfcuj (mgr.14201) 486 : cluster [DBG] pgmap v296: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 294 B/s rd, 589 B/s wr, 0 op/s 2026-03-07T08:03:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:28 vm09 bash[21220]: audit 2026-03-07T07:03:27.698734+0000 mon.vm07 (mon.0) 937 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:28 vm09 bash[21220]: audit 2026-03-07T07:03:27.698734+0000 mon.vm07 (mon.0) 937 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:28 vm09 bash[21220]: audit 2026-03-07T07:03:27.708354+0000 mon.vm07 (mon.0) 938 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:28 vm09 bash[21220]: audit 2026-03-07T07:03:27.708354+0000 mon.vm07 (mon.0) 938 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:28 vm09 bash[21220]: audit 2026-03-07T07:03:27.744417+0000 mon.vm07 (mon.0) 939 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:28 vm09 bash[21220]: audit 2026-03-07T07:03:27.744417+0000 mon.vm07 (mon.0) 939 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:28 vm09 bash[21220]: audit 2026-03-07T07:03:28.145060+0000 mon.vm07 (mon.0) 940 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:28 vm09 bash[21220]: audit 2026-03-07T07:03:28.145060+0000 mon.vm07 (mon.0) 940 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:28 vm09 bash[21220]: audit 2026-03-07T07:03:28.151728+0000 mon.vm07 (mon.0) 941 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:28 vm09 bash[21220]: audit 2026-03-07T07:03:28.151728+0000 mon.vm07 (mon.0) 941 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:29.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:28 vm07 bash[17031]: audit 2026-03-07T07:03:27.698734+0000 mon.vm07 (mon.0) 937 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:29.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:28 vm07 bash[17031]: audit 2026-03-07T07:03:27.698734+0000 mon.vm07 (mon.0) 937 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:29.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:28 vm07 bash[17031]: audit 2026-03-07T07:03:27.708354+0000 mon.vm07 (mon.0) 938 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:29.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:28 vm07 bash[17031]: audit 2026-03-07T07:03:27.708354+0000 mon.vm07 (mon.0) 938 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:29.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:28 vm07 bash[17031]: audit 2026-03-07T07:03:27.744417+0000 mon.vm07 (mon.0) 939 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:29.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:28 vm07 bash[17031]: audit 2026-03-07T07:03:27.744417+0000 mon.vm07 (mon.0) 939 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:29.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:28 vm07 bash[17031]: audit 2026-03-07T07:03:28.145060+0000 mon.vm07 (mon.0) 940 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:29.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:28 vm07 bash[17031]: audit 2026-03-07T07:03:28.145060+0000 mon.vm07 (mon.0) 940 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:29.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:28 vm07 bash[17031]: audit 2026-03-07T07:03:28.151728+0000 mon.vm07 (mon.0) 941 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:29.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:28 vm07 bash[17031]: audit 2026-03-07T07:03:28.151728+0000 mon.vm07 (mon.0) 941 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:30.040 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.bgfaka to start 2026-03-07T08:03:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:29 vm09 bash[21220]: cluster 2026-03-07T07:03:28.762981+0000 mgr.vm07.yrfcuj (mgr.14201) 487 : cluster [DBG] pgmap v297: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 3.3 KiB/s rd, 0 B/s wr, 4 op/s 2026-03-07T08:03:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:29 vm09 bash[21220]: cluster 2026-03-07T07:03:28.762981+0000 mgr.vm07.yrfcuj (mgr.14201) 487 : cluster [DBG] pgmap v297: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 3.3 KiB/s rd, 0 B/s wr, 4 op/s 2026-03-07T08:03:30.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:29 vm07 bash[17031]: cluster 2026-03-07T07:03:28.762981+0000 mgr.vm07.yrfcuj (mgr.14201) 487 : cluster [DBG] pgmap v297: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 3.3 KiB/s rd, 0 B/s wr, 4 op/s 2026-03-07T08:03:30.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:29 vm07 bash[17031]: cluster 2026-03-07T07:03:28.762981+0000 mgr.vm07.yrfcuj (mgr.14201) 487 : cluster [DBG] pgmap v297: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 3.3 KiB/s rd, 0 B/s wr, 4 op/s 2026-03-07T08:03:30.188 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:03:30.188 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 error 22s ago 5m - - 2026-03-07T08:03:30.188 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 22s ago 5m 114M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:03:30.188 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (5m) 88s ago 5m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:03:30.188 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (5m) 88s ago 5m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:03:30.369 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:03:30.369 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:03:30.369 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm07.bgfaka on vm07 is in error state 2026-03-07T08:03:31.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:30 vm09 bash[21220]: audit 2026-03-07T07:03:30.025005+0000 mgr.vm07.yrfcuj (mgr.14201) 488 : audit [DBG] from='client.15328 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:31.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:30 vm09 bash[21220]: audit 2026-03-07T07:03:30.025005+0000 mgr.vm07.yrfcuj (mgr.14201) 488 : audit [DBG] from='client.15328 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:31.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:30 vm09 bash[21220]: audit 2026-03-07T07:03:30.185456+0000 mgr.vm07.yrfcuj (mgr.14201) 489 : audit [DBG] from='client.15332 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:31.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:30 vm09 bash[21220]: audit 2026-03-07T07:03:30.185456+0000 mgr.vm07.yrfcuj (mgr.14201) 489 : audit [DBG] from='client.15332 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:31.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:30 vm09 bash[21220]: audit 2026-03-07T07:03:30.368658+0000 mon.vm09 (mon.1) 35 : audit [DBG] from='client.? 192.168.123.107:0/2984452406' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:31.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:30 vm09 bash[21220]: audit 2026-03-07T07:03:30.368658+0000 mon.vm09 (mon.1) 35 : audit [DBG] from='client.? 192.168.123.107:0/2984452406' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:31.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:30 vm07 bash[17031]: audit 2026-03-07T07:03:30.025005+0000 mgr.vm07.yrfcuj (mgr.14201) 488 : audit [DBG] from='client.15328 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:31.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:30 vm07 bash[17031]: audit 2026-03-07T07:03:30.025005+0000 mgr.vm07.yrfcuj (mgr.14201) 488 : audit [DBG] from='client.15328 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:31.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:30 vm07 bash[17031]: audit 2026-03-07T07:03:30.185456+0000 mgr.vm07.yrfcuj (mgr.14201) 489 : audit [DBG] from='client.15332 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:31.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:30 vm07 bash[17031]: audit 2026-03-07T07:03:30.185456+0000 mgr.vm07.yrfcuj (mgr.14201) 489 : audit [DBG] from='client.15332 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:31.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:30 vm07 bash[17031]: audit 2026-03-07T07:03:30.368658+0000 mon.vm09 (mon.1) 35 : audit [DBG] from='client.? 192.168.123.107:0/2984452406' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:31.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:30 vm07 bash[17031]: audit 2026-03-07T07:03:30.368658+0000 mon.vm09 (mon.1) 35 : audit [DBG] from='client.? 192.168.123.107:0/2984452406' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:32.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:31 vm09 bash[21220]: cluster 2026-03-07T07:03:30.763353+0000 mgr.vm07.yrfcuj (mgr.14201) 490 : cluster [DBG] pgmap v298: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 9.8 KiB/s rd, 0 B/s wr, 15 op/s 2026-03-07T08:03:32.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:31 vm09 bash[21220]: cluster 2026-03-07T07:03:30.763353+0000 mgr.vm07.yrfcuj (mgr.14201) 490 : cluster [DBG] pgmap v298: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 9.8 KiB/s rd, 0 B/s wr, 15 op/s 2026-03-07T08:03:32.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:31 vm07 bash[17031]: cluster 2026-03-07T07:03:30.763353+0000 mgr.vm07.yrfcuj (mgr.14201) 490 : cluster [DBG] pgmap v298: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 9.8 KiB/s rd, 0 B/s wr, 15 op/s 2026-03-07T08:03:32.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:31 vm07 bash[17031]: cluster 2026-03-07T07:03:30.763353+0000 mgr.vm07.yrfcuj (mgr.14201) 490 : cluster [DBG] pgmap v298: 129 pgs: 129 active+clean; 454 KiB data, 230 MiB used, 160 GiB / 160 GiB avail; 9.8 KiB/s rd, 0 B/s wr, 15 op/s 2026-03-07T08:03:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:33 vm09 bash[21220]: cluster 2026-03-07T07:03:32.763915+0000 mgr.vm07.yrfcuj (mgr.14201) 491 : cluster [DBG] pgmap v299: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 0 B/s wr, 82 op/s 2026-03-07T08:03:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:33 vm09 bash[21220]: cluster 2026-03-07T07:03:32.763915+0000 mgr.vm07.yrfcuj (mgr.14201) 491 : cluster [DBG] pgmap v299: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 0 B/s wr, 82 op/s 2026-03-07T08:03:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:33 vm09 bash[21220]: audit 2026-03-07T07:03:33.690119+0000 mon.vm07 (mon.0) 942 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:33 vm09 bash[21220]: audit 2026-03-07T07:03:33.690119+0000 mon.vm07 (mon.0) 942 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:33 vm09 bash[21220]: audit 2026-03-07T07:03:33.695778+0000 mon.vm07 (mon.0) 943 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:33 vm09 bash[21220]: audit 2026-03-07T07:03:33.695778+0000 mon.vm07 (mon.0) 943 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:33 vm09 bash[21220]: audit 2026-03-07T07:03:33.696988+0000 mon.vm07 (mon.0) 944 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:33 vm09 bash[21220]: audit 2026-03-07T07:03:33.696988+0000 mon.vm07 (mon.0) 944 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:33 vm09 bash[21220]: audit 2026-03-07T07:03:33.697841+0000 mon.vm07 (mon.0) 945 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:33 vm09 bash[21220]: audit 2026-03-07T07:03:33.697841+0000 mon.vm07 (mon.0) 945 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:33 vm09 bash[21220]: audit 2026-03-07T07:03:33.702406+0000 mon.vm07 (mon.0) 946 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:33 vm09 bash[21220]: audit 2026-03-07T07:03:33.702406+0000 mon.vm07 (mon.0) 946 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:33 vm09 bash[21220]: audit 2026-03-07T07:03:33.704573+0000 mon.vm07 (mon.0) 947 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:33 vm09 bash[21220]: audit 2026-03-07T07:03:33.704573+0000 mon.vm07 (mon.0) 947 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:34.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:33 vm07 bash[17031]: cluster 2026-03-07T07:03:32.763915+0000 mgr.vm07.yrfcuj (mgr.14201) 491 : cluster [DBG] pgmap v299: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 0 B/s wr, 82 op/s 2026-03-07T08:03:34.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:33 vm07 bash[17031]: cluster 2026-03-07T07:03:32.763915+0000 mgr.vm07.yrfcuj (mgr.14201) 491 : cluster [DBG] pgmap v299: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 0 B/s wr, 82 op/s 2026-03-07T08:03:34.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:33 vm07 bash[17031]: audit 2026-03-07T07:03:33.690119+0000 mon.vm07 (mon.0) 942 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:34.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:33 vm07 bash[17031]: audit 2026-03-07T07:03:33.690119+0000 mon.vm07 (mon.0) 942 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:34.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:33 vm07 bash[17031]: audit 2026-03-07T07:03:33.695778+0000 mon.vm07 (mon.0) 943 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:34.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:33 vm07 bash[17031]: audit 2026-03-07T07:03:33.695778+0000 mon.vm07 (mon.0) 943 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:34.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:33 vm07 bash[17031]: audit 2026-03-07T07:03:33.696988+0000 mon.vm07 (mon.0) 944 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:34.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:33 vm07 bash[17031]: audit 2026-03-07T07:03:33.696988+0000 mon.vm07 (mon.0) 944 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:34.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:33 vm07 bash[17031]: audit 2026-03-07T07:03:33.697841+0000 mon.vm07 (mon.0) 945 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:34.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:33 vm07 bash[17031]: audit 2026-03-07T07:03:33.697841+0000 mon.vm07 (mon.0) 945 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:34.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:33 vm07 bash[17031]: audit 2026-03-07T07:03:33.702406+0000 mon.vm07 (mon.0) 946 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:34.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:33 vm07 bash[17031]: audit 2026-03-07T07:03:33.702406+0000 mon.vm07 (mon.0) 946 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:34.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:33 vm07 bash[17031]: audit 2026-03-07T07:03:33.704573+0000 mon.vm07 (mon.0) 947 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:34.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:33 vm07 bash[17031]: audit 2026-03-07T07:03:33.704573+0000 mon.vm07 (mon.0) 947 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:35.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:34 vm09 bash[21220]: cluster 2026-03-07T07:03:33.699124+0000 mgr.vm07.yrfcuj (mgr.14201) 492 : cluster [DBG] pgmap v300: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 58 KiB/s rd, 0 B/s wr, 90 op/s 2026-03-07T08:03:35.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:34 vm09 bash[21220]: cluster 2026-03-07T07:03:33.699124+0000 mgr.vm07.yrfcuj (mgr.14201) 492 : cluster [DBG] pgmap v300: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 58 KiB/s rd, 0 B/s wr, 90 op/s 2026-03-07T08:03:35.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:34 vm09 bash[21220]: cluster 2026-03-07T07:03:34.694259+0000 mon.vm07 (mon.0) 948 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-07T08:03:35.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:34 vm09 bash[21220]: cluster 2026-03-07T07:03:34.694259+0000 mon.vm07 (mon.0) 948 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-07T08:03:35.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:34 vm09 bash[21220]: cluster 2026-03-07T07:03:34.694277+0000 mon.vm07 (mon.0) 949 : cluster [INF] Cluster is now healthy 2026-03-07T08:03:35.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:34 vm09 bash[21220]: cluster 2026-03-07T07:03:34.694277+0000 mon.vm07 (mon.0) 949 : cluster [INF] Cluster is now healthy 2026-03-07T08:03:35.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:34 vm07 bash[17031]: cluster 2026-03-07T07:03:33.699124+0000 mgr.vm07.yrfcuj (mgr.14201) 492 : cluster [DBG] pgmap v300: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 58 KiB/s rd, 0 B/s wr, 90 op/s 2026-03-07T08:03:35.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:34 vm07 bash[17031]: cluster 2026-03-07T07:03:33.699124+0000 mgr.vm07.yrfcuj (mgr.14201) 492 : cluster [DBG] pgmap v300: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 58 KiB/s rd, 0 B/s wr, 90 op/s 2026-03-07T08:03:35.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:34 vm07 bash[17031]: cluster 2026-03-07T07:03:34.694259+0000 mon.vm07 (mon.0) 948 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-07T08:03:35.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:34 vm07 bash[17031]: cluster 2026-03-07T07:03:34.694259+0000 mon.vm07 (mon.0) 948 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-07T08:03:35.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:34 vm07 bash[17031]: cluster 2026-03-07T07:03:34.694277+0000 mon.vm07 (mon.0) 949 : cluster [INF] Cluster is now healthy 2026-03-07T08:03:35.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:34 vm07 bash[17031]: cluster 2026-03-07T07:03:34.694277+0000 mon.vm07 (mon.0) 949 : cluster [INF] Cluster is now healthy 2026-03-07T08:03:35.552 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (7s) 1s ago 5m 89.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:03:35.723 INFO:teuthology.orchestra.run.vm07.stdout:Scheduled to stop rgw.foo.vm07.lrxyjy on host 'vm07' 2026-03-07T08:03:35.976 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.lrxyjy to stop 2026-03-07T08:03:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:35 vm09 bash[21220]: audit 2026-03-07T07:03:35.538317+0000 mgr.vm07.yrfcuj (mgr.14201) 493 : audit [DBG] from='client.15340 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:35 vm09 bash[21220]: audit 2026-03-07T07:03:35.538317+0000 mgr.vm07.yrfcuj (mgr.14201) 493 : audit [DBG] from='client.15340 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:35 vm09 bash[21220]: audit 2026-03-07T07:03:35.714319+0000 mon.vm07 (mon.0) 950 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:35 vm09 bash[21220]: audit 2026-03-07T07:03:35.714319+0000 mon.vm07 (mon.0) 950 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:35 vm09 bash[21220]: audit 2026-03-07T07:03:35.720548+0000 mon.vm07 (mon.0) 951 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:35 vm09 bash[21220]: audit 2026-03-07T07:03:35.720548+0000 mon.vm07 (mon.0) 951 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:35 vm09 bash[21220]: audit 2026-03-07T07:03:35.721932+0000 mon.vm07 (mon.0) 952 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:35 vm09 bash[21220]: audit 2026-03-07T07:03:35.721932+0000 mon.vm07 (mon.0) 952 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:35 vm09 bash[21220]: audit 2026-03-07T07:03:35.723877+0000 mon.vm07 (mon.0) 953 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:35 vm09 bash[21220]: audit 2026-03-07T07:03:35.723877+0000 mon.vm07 (mon.0) 953 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:35 vm09 bash[21220]: audit 2026-03-07T07:03:35.726069+0000 mon.vm07 (mon.0) 954 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:35 vm09 bash[21220]: audit 2026-03-07T07:03:35.726069+0000 mon.vm07 (mon.0) 954 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:35 vm09 bash[21220]: audit 2026-03-07T07:03:35.748209+0000 mon.vm07 (mon.0) 955 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:35 vm09 bash[21220]: audit 2026-03-07T07:03:35.748209+0000 mon.vm07 (mon.0) 955 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:35 vm09 bash[21220]: audit 2026-03-07T07:03:35.749929+0000 mon.vm07 (mon.0) 956 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:35 vm09 bash[21220]: audit 2026-03-07T07:03:35.749929+0000 mon.vm07 (mon.0) 956 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:36.138 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:03:36.138 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (8s) 2s ago 5m 89.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:03:36.138 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 2s ago 5m 116M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:03:36.138 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (5m) 94s ago 5m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:03:36.138 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (5m) 94s ago 5m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:03:36.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:35 vm07 bash[17031]: audit 2026-03-07T07:03:35.538317+0000 mgr.vm07.yrfcuj (mgr.14201) 493 : audit [DBG] from='client.15340 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:36.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:35 vm07 bash[17031]: audit 2026-03-07T07:03:35.538317+0000 mgr.vm07.yrfcuj (mgr.14201) 493 : audit [DBG] from='client.15340 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:36.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:35 vm07 bash[17031]: audit 2026-03-07T07:03:35.714319+0000 mon.vm07 (mon.0) 950 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:36.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:35 vm07 bash[17031]: audit 2026-03-07T07:03:35.714319+0000 mon.vm07 (mon.0) 950 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:36.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:35 vm07 bash[17031]: audit 2026-03-07T07:03:35.720548+0000 mon.vm07 (mon.0) 951 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:36.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:35 vm07 bash[17031]: audit 2026-03-07T07:03:35.720548+0000 mon.vm07 (mon.0) 951 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:36.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:35 vm07 bash[17031]: audit 2026-03-07T07:03:35.721932+0000 mon.vm07 (mon.0) 952 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:36.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:35 vm07 bash[17031]: audit 2026-03-07T07:03:35.721932+0000 mon.vm07 (mon.0) 952 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:36.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:35 vm07 bash[17031]: audit 2026-03-07T07:03:35.723877+0000 mon.vm07 (mon.0) 953 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:36.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:35 vm07 bash[17031]: audit 2026-03-07T07:03:35.723877+0000 mon.vm07 (mon.0) 953 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:36.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:35 vm07 bash[17031]: audit 2026-03-07T07:03:35.726069+0000 mon.vm07 (mon.0) 954 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:36.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:35 vm07 bash[17031]: audit 2026-03-07T07:03:35.726069+0000 mon.vm07 (mon.0) 954 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:36.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:35 vm07 bash[17031]: audit 2026-03-07T07:03:35.748209+0000 mon.vm07 (mon.0) 955 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:36.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:35 vm07 bash[17031]: audit 2026-03-07T07:03:35.748209+0000 mon.vm07 (mon.0) 955 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:36.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:35 vm07 bash[17031]: audit 2026-03-07T07:03:35.749929+0000 mon.vm07 (mon.0) 956 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:36.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:35 vm07 bash[17031]: audit 2026-03-07T07:03:35.749929+0000 mon.vm07 (mon.0) 956 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:36.330 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:03:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:36 vm09 bash[21220]: cluster 2026-03-07T07:03:35.699465+0000 mgr.vm07.yrfcuj (mgr.14201) 494 : cluster [DBG] pgmap v301: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 58 KiB/s rd, 0 B/s wr, 90 op/s 2026-03-07T08:03:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:36 vm09 bash[21220]: cluster 2026-03-07T07:03:35.699465+0000 mgr.vm07.yrfcuj (mgr.14201) 494 : cluster [DBG] pgmap v301: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 58 KiB/s rd, 0 B/s wr, 90 op/s 2026-03-07T08:03:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:36 vm09 bash[21220]: audit 2026-03-07T07:03:35.707737+0000 mgr.vm07.yrfcuj (mgr.14201) 495 : audit [DBG] from='client.15344 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "rgw.foo.vm07.lrxyjy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:36 vm09 bash[21220]: audit 2026-03-07T07:03:35.707737+0000 mgr.vm07.yrfcuj (mgr.14201) 495 : audit [DBG] from='client.15344 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "rgw.foo.vm07.lrxyjy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:36 vm09 bash[21220]: cephadm 2026-03-07T07:03:35.708196+0000 mgr.vm07.yrfcuj (mgr.14201) 496 : cephadm [INF] Schedule stop daemon rgw.foo.vm07.lrxyjy 2026-03-07T08:03:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:36 vm09 bash[21220]: cephadm 2026-03-07T07:03:35.708196+0000 mgr.vm07.yrfcuj (mgr.14201) 496 : cephadm [INF] Schedule stop daemon rgw.foo.vm07.lrxyjy 2026-03-07T08:03:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:36 vm09 bash[21220]: audit 2026-03-07T07:03:35.955815+0000 mgr.vm07.yrfcuj (mgr.14201) 497 : audit [DBG] from='client.24871 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:36 vm09 bash[21220]: audit 2026-03-07T07:03:35.955815+0000 mgr.vm07.yrfcuj (mgr.14201) 497 : audit [DBG] from='client.24871 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:36 vm09 bash[21220]: audit 2026-03-07T07:03:36.135554+0000 mgr.vm07.yrfcuj (mgr.14201) 498 : audit [DBG] from='client.15352 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:36 vm09 bash[21220]: audit 2026-03-07T07:03:36.135554+0000 mgr.vm07.yrfcuj (mgr.14201) 498 : audit [DBG] from='client.15352 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:36 vm09 bash[21220]: audit 2026-03-07T07:03:36.329919+0000 mon.vm09 (mon.1) 36 : audit [DBG] from='client.? 192.168.123.107:0/2778884546' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:36 vm09 bash[21220]: audit 2026-03-07T07:03:36.329919+0000 mon.vm09 (mon.1) 36 : audit [DBG] from='client.? 192.168.123.107:0/2778884546' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:37.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:36 vm07 bash[17031]: cluster 2026-03-07T07:03:35.699465+0000 mgr.vm07.yrfcuj (mgr.14201) 494 : cluster [DBG] pgmap v301: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 58 KiB/s rd, 0 B/s wr, 90 op/s 2026-03-07T08:03:37.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:36 vm07 bash[17031]: cluster 2026-03-07T07:03:35.699465+0000 mgr.vm07.yrfcuj (mgr.14201) 494 : cluster [DBG] pgmap v301: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 58 KiB/s rd, 0 B/s wr, 90 op/s 2026-03-07T08:03:37.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:36 vm07 bash[17031]: audit 2026-03-07T07:03:35.707737+0000 mgr.vm07.yrfcuj (mgr.14201) 495 : audit [DBG] from='client.15344 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "rgw.foo.vm07.lrxyjy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:37.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:36 vm07 bash[17031]: audit 2026-03-07T07:03:35.707737+0000 mgr.vm07.yrfcuj (mgr.14201) 495 : audit [DBG] from='client.15344 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "rgw.foo.vm07.lrxyjy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:37.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:36 vm07 bash[17031]: cephadm 2026-03-07T07:03:35.708196+0000 mgr.vm07.yrfcuj (mgr.14201) 496 : cephadm [INF] Schedule stop daemon rgw.foo.vm07.lrxyjy 2026-03-07T08:03:37.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:36 vm07 bash[17031]: cephadm 2026-03-07T07:03:35.708196+0000 mgr.vm07.yrfcuj (mgr.14201) 496 : cephadm [INF] Schedule stop daemon rgw.foo.vm07.lrxyjy 2026-03-07T08:03:37.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:36 vm07 bash[17031]: audit 2026-03-07T07:03:35.955815+0000 mgr.vm07.yrfcuj (mgr.14201) 497 : audit [DBG] from='client.24871 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:37.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:36 vm07 bash[17031]: audit 2026-03-07T07:03:35.955815+0000 mgr.vm07.yrfcuj (mgr.14201) 497 : audit [DBG] from='client.24871 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:37.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:36 vm07 bash[17031]: audit 2026-03-07T07:03:36.135554+0000 mgr.vm07.yrfcuj (mgr.14201) 498 : audit [DBG] from='client.15352 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:37.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:36 vm07 bash[17031]: audit 2026-03-07T07:03:36.135554+0000 mgr.vm07.yrfcuj (mgr.14201) 498 : audit [DBG] from='client.15352 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:37.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:36 vm07 bash[17031]: audit 2026-03-07T07:03:36.329919+0000 mon.vm09 (mon.1) 36 : audit [DBG] from='client.? 192.168.123.107:0/2778884546' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:37.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:36 vm07 bash[17031]: audit 2026-03-07T07:03:36.329919+0000 mon.vm09 (mon.1) 36 : audit [DBG] from='client.? 192.168.123.107:0/2778884546' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:39.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:38 vm09 bash[21220]: cluster 2026-03-07T07:03:37.699995+0000 mgr.vm07.yrfcuj (mgr.14201) 499 : cluster [DBG] pgmap v302: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 58 KiB/s rd, 0 B/s wr, 90 op/s 2026-03-07T08:03:39.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:38 vm09 bash[21220]: cluster 2026-03-07T07:03:37.699995+0000 mgr.vm07.yrfcuj (mgr.14201) 499 : cluster [DBG] pgmap v302: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 58 KiB/s rd, 0 B/s wr, 90 op/s 2026-03-07T08:03:39.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:38 vm07 bash[17031]: cluster 2026-03-07T07:03:37.699995+0000 mgr.vm07.yrfcuj (mgr.14201) 499 : cluster [DBG] pgmap v302: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 58 KiB/s rd, 0 B/s wr, 90 op/s 2026-03-07T08:03:39.153 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:38 vm07 bash[17031]: cluster 2026-03-07T07:03:37.699995+0000 mgr.vm07.yrfcuj (mgr.14201) 499 : cluster [DBG] pgmap v302: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 58 KiB/s rd, 0 B/s wr, 90 op/s 2026-03-07T08:03:41.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:41 vm07 bash[17031]: cluster 2026-03-07T07:03:39.700355+0000 mgr.vm07.yrfcuj (mgr.14201) 500 : cluster [DBG] pgmap v303: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 55 KiB/s rd, 0 B/s wr, 85 op/s 2026-03-07T08:03:41.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:41 vm07 bash[17031]: cluster 2026-03-07T07:03:39.700355+0000 mgr.vm07.yrfcuj (mgr.14201) 500 : cluster [DBG] pgmap v303: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 55 KiB/s rd, 0 B/s wr, 85 op/s 2026-03-07T08:03:41.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:41 vm07 bash[17031]: audit 2026-03-07T07:03:40.138967+0000 mon.vm07 (mon.0) 957 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:41.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:41 vm07 bash[17031]: audit 2026-03-07T07:03:40.138967+0000 mon.vm07 (mon.0) 957 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:41.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:41 vm07 bash[17031]: audit 2026-03-07T07:03:40.139721+0000 mon.vm07 (mon.0) 958 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:03:41.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:41 vm07 bash[17031]: audit 2026-03-07T07:03:40.139721+0000 mon.vm07 (mon.0) 958 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:03:41.501 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.lrxyjy to stop 2026-03-07T08:03:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:41 vm09 bash[21220]: cluster 2026-03-07T07:03:39.700355+0000 mgr.vm07.yrfcuj (mgr.14201) 500 : cluster [DBG] pgmap v303: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 55 KiB/s rd, 0 B/s wr, 85 op/s 2026-03-07T08:03:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:41 vm09 bash[21220]: cluster 2026-03-07T07:03:39.700355+0000 mgr.vm07.yrfcuj (mgr.14201) 500 : cluster [DBG] pgmap v303: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 55 KiB/s rd, 0 B/s wr, 85 op/s 2026-03-07T08:03:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:41 vm09 bash[21220]: audit 2026-03-07T07:03:40.138967+0000 mon.vm07 (mon.0) 957 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:41 vm09 bash[21220]: audit 2026-03-07T07:03:40.138967+0000 mon.vm07 (mon.0) 957 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:41 vm09 bash[21220]: audit 2026-03-07T07:03:40.139721+0000 mon.vm07 (mon.0) 958 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:03:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:41 vm09 bash[21220]: audit 2026-03-07T07:03:40.139721+0000 mon.vm07 (mon.0) 958 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:03:41.653 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:03:41.653 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (13s) 7s ago 6m 89.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:03:41.653 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (6m) 7s ago 6m 116M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:03:41.653 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (6m) 99s ago 6m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:03:41.653 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (6m) 99s ago 6m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:03:41.841 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:03:42.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:42 vm07 bash[17031]: audit 2026-03-07T07:03:41.485311+0000 mgr.vm07.yrfcuj (mgr.14201) 501 : audit [DBG] from='client.15360 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:42.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:42 vm07 bash[17031]: audit 2026-03-07T07:03:41.485311+0000 mgr.vm07.yrfcuj (mgr.14201) 501 : audit [DBG] from='client.15360 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:42.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:42 vm07 bash[17031]: audit 2026-03-07T07:03:41.840410+0000 mon.vm09 (mon.1) 37 : audit [DBG] from='client.? 192.168.123.107:0/3502586628' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:42.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:42 vm07 bash[17031]: audit 2026-03-07T07:03:41.840410+0000 mon.vm09 (mon.1) 37 : audit [DBG] from='client.? 192.168.123.107:0/3502586628' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:42.614 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:42 vm09 bash[21220]: audit 2026-03-07T07:03:41.485311+0000 mgr.vm07.yrfcuj (mgr.14201) 501 : audit [DBG] from='client.15360 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:42.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:42 vm09 bash[21220]: audit 2026-03-07T07:03:41.485311+0000 mgr.vm07.yrfcuj (mgr.14201) 501 : audit [DBG] from='client.15360 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:42.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:42 vm09 bash[21220]: audit 2026-03-07T07:03:41.840410+0000 mon.vm09 (mon.1) 37 : audit [DBG] from='client.? 192.168.123.107:0/3502586628' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:42.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:42 vm09 bash[21220]: audit 2026-03-07T07:03:41.840410+0000 mon.vm09 (mon.1) 37 : audit [DBG] from='client.? 192.168.123.107:0/3502586628' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:43.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:43 vm09 bash[21220]: audit 2026-03-07T07:03:41.650644+0000 mgr.vm07.yrfcuj (mgr.14201) 502 : audit [DBG] from='client.24887 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:43.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:43 vm09 bash[21220]: audit 2026-03-07T07:03:41.650644+0000 mgr.vm07.yrfcuj (mgr.14201) 502 : audit [DBG] from='client.24887 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:43.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:43 vm09 bash[21220]: cluster 2026-03-07T07:03:41.700729+0000 mgr.vm07.yrfcuj (mgr.14201) 503 : cluster [DBG] pgmap v304: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 47 KiB/s rd, 187 B/s wr, 73 op/s 2026-03-07T08:03:43.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:43 vm09 bash[21220]: cluster 2026-03-07T07:03:41.700729+0000 mgr.vm07.yrfcuj (mgr.14201) 503 : cluster [DBG] pgmap v304: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 47 KiB/s rd, 187 B/s wr, 73 op/s 2026-03-07T08:03:43.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:43 vm07 bash[17031]: audit 2026-03-07T07:03:41.650644+0000 mgr.vm07.yrfcuj (mgr.14201) 502 : audit [DBG] from='client.24887 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:43.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:43 vm07 bash[17031]: audit 2026-03-07T07:03:41.650644+0000 mgr.vm07.yrfcuj (mgr.14201) 502 : audit [DBG] from='client.24887 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:43.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:43 vm07 bash[17031]: cluster 2026-03-07T07:03:41.700729+0000 mgr.vm07.yrfcuj (mgr.14201) 503 : cluster [DBG] pgmap v304: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 47 KiB/s rd, 187 B/s wr, 73 op/s 2026-03-07T08:03:43.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:43 vm07 bash[17031]: cluster 2026-03-07T07:03:41.700729+0000 mgr.vm07.yrfcuj (mgr.14201) 503 : cluster [DBG] pgmap v304: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 47 KiB/s rd, 187 B/s wr, 73 op/s 2026-03-07T08:03:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:45 vm09 bash[21220]: cluster 2026-03-07T07:03:43.701158+0000 mgr.vm07.yrfcuj (mgr.14201) 504 : cluster [DBG] pgmap v305: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 93 B/s rd, 187 B/s wr, 0 op/s 2026-03-07T08:03:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:45 vm09 bash[21220]: cluster 2026-03-07T07:03:43.701158+0000 mgr.vm07.yrfcuj (mgr.14201) 504 : cluster [DBG] pgmap v305: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 93 B/s rd, 187 B/s wr, 0 op/s 2026-03-07T08:03:45.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:45 vm07 bash[17031]: cluster 2026-03-07T07:03:43.701158+0000 mgr.vm07.yrfcuj (mgr.14201) 504 : cluster [DBG] pgmap v305: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 93 B/s rd, 187 B/s wr, 0 op/s 2026-03-07T08:03:45.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:45 vm07 bash[17031]: cluster 2026-03-07T07:03:43.701158+0000 mgr.vm07.yrfcuj (mgr.14201) 504 : cluster [DBG] pgmap v305: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 93 B/s rd, 187 B/s wr, 0 op/s 2026-03-07T08:03:47.018 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.lrxyjy to stop 2026-03-07T08:03:47.183 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:03:47.183 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (19s) 13s ago 6m 89.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:03:47.183 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (6m) 13s ago 6m 116M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:03:47.183 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (6m) 105s ago 6m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:03:47.183 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (6m) 105s ago 6m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:03:47.373 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:03:47.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:47 vm09 bash[21220]: cluster 2026-03-07T07:03:45.701533+0000 mgr.vm07.yrfcuj (mgr.14201) 505 : cluster [DBG] pgmap v306: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:03:47.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:47 vm09 bash[21220]: cluster 2026-03-07T07:03:45.701533+0000 mgr.vm07.yrfcuj (mgr.14201) 505 : cluster [DBG] pgmap v306: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:03:47.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:47 vm07 bash[17031]: cluster 2026-03-07T07:03:45.701533+0000 mgr.vm07.yrfcuj (mgr.14201) 505 : cluster [DBG] pgmap v306: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:03:47.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:47 vm07 bash[17031]: cluster 2026-03-07T07:03:45.701533+0000 mgr.vm07.yrfcuj (mgr.14201) 505 : cluster [DBG] pgmap v306: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:03:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:48 vm09 bash[21220]: audit 2026-03-07T07:03:47.002693+0000 mgr.vm07.yrfcuj (mgr.14201) 506 : audit [DBG] from='client.15372 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:48 vm09 bash[21220]: audit 2026-03-07T07:03:47.002693+0000 mgr.vm07.yrfcuj (mgr.14201) 506 : audit [DBG] from='client.15372 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:48 vm09 bash[21220]: audit 2026-03-07T07:03:47.181356+0000 mgr.vm07.yrfcuj (mgr.14201) 507 : audit [DBG] from='client.15376 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:48 vm09 bash[21220]: audit 2026-03-07T07:03:47.181356+0000 mgr.vm07.yrfcuj (mgr.14201) 507 : audit [DBG] from='client.15376 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:48 vm09 bash[21220]: audit 2026-03-07T07:03:47.372157+0000 mon.vm09 (mon.1) 38 : audit [DBG] from='client.? 192.168.123.107:0/1694347982' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:48 vm09 bash[21220]: audit 2026-03-07T07:03:47.372157+0000 mon.vm09 (mon.1) 38 : audit [DBG] from='client.? 192.168.123.107:0/1694347982' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:48.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:48 vm07 bash[17031]: audit 2026-03-07T07:03:47.002693+0000 mgr.vm07.yrfcuj (mgr.14201) 506 : audit [DBG] from='client.15372 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:48.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:48 vm07 bash[17031]: audit 2026-03-07T07:03:47.002693+0000 mgr.vm07.yrfcuj (mgr.14201) 506 : audit [DBG] from='client.15372 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:48.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:48 vm07 bash[17031]: audit 2026-03-07T07:03:47.181356+0000 mgr.vm07.yrfcuj (mgr.14201) 507 : audit [DBG] from='client.15376 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:48.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:48 vm07 bash[17031]: audit 2026-03-07T07:03:47.181356+0000 mgr.vm07.yrfcuj (mgr.14201) 507 : audit [DBG] from='client.15376 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:48.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:48 vm07 bash[17031]: audit 2026-03-07T07:03:47.372157+0000 mon.vm09 (mon.1) 38 : audit [DBG] from='client.? 192.168.123.107:0/1694347982' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:48.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:48 vm07 bash[17031]: audit 2026-03-07T07:03:47.372157+0000 mon.vm09 (mon.1) 38 : audit [DBG] from='client.? 192.168.123.107:0/1694347982' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:49.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:49 vm09 bash[21220]: cluster 2026-03-07T07:03:47.701987+0000 mgr.vm07.yrfcuj (mgr.14201) 508 : cluster [DBG] pgmap v307: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:03:49.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:49 vm09 bash[21220]: cluster 2026-03-07T07:03:47.701987+0000 mgr.vm07.yrfcuj (mgr.14201) 508 : cluster [DBG] pgmap v307: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:03:49.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:49 vm07 bash[17031]: cluster 2026-03-07T07:03:47.701987+0000 mgr.vm07.yrfcuj (mgr.14201) 508 : cluster [DBG] pgmap v307: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:03:49.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:49 vm07 bash[17031]: cluster 2026-03-07T07:03:47.701987+0000 mgr.vm07.yrfcuj (mgr.14201) 508 : cluster [DBG] pgmap v307: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:03:51.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:50 vm09 bash[21220]: cluster 2026-03-07T07:03:49.702320+0000 mgr.vm07.yrfcuj (mgr.14201) 509 : cluster [DBG] pgmap v308: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:03:51.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:50 vm09 bash[21220]: cluster 2026-03-07T07:03:49.702320+0000 mgr.vm07.yrfcuj (mgr.14201) 509 : cluster [DBG] pgmap v308: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:03:51.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:50 vm09 bash[21220]: audit 2026-03-07T07:03:49.910936+0000 mon.vm07 (mon.0) 959 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:51.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:50 vm09 bash[21220]: audit 2026-03-07T07:03:49.910936+0000 mon.vm07 (mon.0) 959 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:51.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:50 vm09 bash[21220]: audit 2026-03-07T07:03:49.915437+0000 mon.vm07 (mon.0) 960 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:51.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:50 vm09 bash[21220]: audit 2026-03-07T07:03:49.915437+0000 mon.vm07 (mon.0) 960 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:51.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:50 vm09 bash[21220]: audit 2026-03-07T07:03:49.943629+0000 mon.vm07 (mon.0) 961 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:51.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:50 vm09 bash[21220]: audit 2026-03-07T07:03:49.943629+0000 mon.vm07 (mon.0) 961 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:51.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:50 vm07 bash[17031]: cluster 2026-03-07T07:03:49.702320+0000 mgr.vm07.yrfcuj (mgr.14201) 509 : cluster [DBG] pgmap v308: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:03:51.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:50 vm07 bash[17031]: cluster 2026-03-07T07:03:49.702320+0000 mgr.vm07.yrfcuj (mgr.14201) 509 : cluster [DBG] pgmap v308: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:03:51.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:50 vm07 bash[17031]: audit 2026-03-07T07:03:49.910936+0000 mon.vm07 (mon.0) 959 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:51.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:50 vm07 bash[17031]: audit 2026-03-07T07:03:49.910936+0000 mon.vm07 (mon.0) 959 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:51.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:50 vm07 bash[17031]: audit 2026-03-07T07:03:49.915437+0000 mon.vm07 (mon.0) 960 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:51.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:50 vm07 bash[17031]: audit 2026-03-07T07:03:49.915437+0000 mon.vm07 (mon.0) 960 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:51.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:50 vm07 bash[17031]: audit 2026-03-07T07:03:49.943629+0000 mon.vm07 (mon.0) 961 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:51.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:50 vm07 bash[17031]: audit 2026-03-07T07:03:49.943629+0000 mon.vm07 (mon.0) 961 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:52.551 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.lrxyjy to stop 2026-03-07T08:03:52.702 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:03:52.702 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (25s) 19s ago 6m 89.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:03:52.702 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (6m) 19s ago 6m 116M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 335df92b516d 2026-03-07T08:03:52.702 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (6m) 110s ago 6m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:03:52.702 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (6m) 110s ago 6m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:03:52.894 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:03:53.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:52 vm09 bash[21220]: cluster 2026-03-07T07:03:51.702680+0000 mgr.vm07.yrfcuj (mgr.14201) 510 : cluster [DBG] pgmap v309: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:03:53.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:52 vm09 bash[21220]: cluster 2026-03-07T07:03:51.702680+0000 mgr.vm07.yrfcuj (mgr.14201) 510 : cluster [DBG] pgmap v309: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:03:53.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:52 vm09 bash[21220]: audit 2026-03-07T07:03:52.536451+0000 mgr.vm07.yrfcuj (mgr.14201) 511 : audit [DBG] from='client.15384 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:53.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:52 vm09 bash[21220]: audit 2026-03-07T07:03:52.536451+0000 mgr.vm07.yrfcuj (mgr.14201) 511 : audit [DBG] from='client.15384 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:53.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:52 vm09 bash[21220]: audit 2026-03-07T07:03:52.894853+0000 mon.vm07 (mon.0) 962 : audit [DBG] from='client.? 192.168.123.107:0/1668099927' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:53.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:52 vm09 bash[21220]: audit 2026-03-07T07:03:52.894853+0000 mon.vm07 (mon.0) 962 : audit [DBG] from='client.? 192.168.123.107:0/1668099927' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:53.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:52 vm07 bash[17031]: cluster 2026-03-07T07:03:51.702680+0000 mgr.vm07.yrfcuj (mgr.14201) 510 : cluster [DBG] pgmap v309: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:03:53.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:52 vm07 bash[17031]: cluster 2026-03-07T07:03:51.702680+0000 mgr.vm07.yrfcuj (mgr.14201) 510 : cluster [DBG] pgmap v309: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:03:53.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:52 vm07 bash[17031]: audit 2026-03-07T07:03:52.536451+0000 mgr.vm07.yrfcuj (mgr.14201) 511 : audit [DBG] from='client.15384 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:53.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:52 vm07 bash[17031]: audit 2026-03-07T07:03:52.536451+0000 mgr.vm07.yrfcuj (mgr.14201) 511 : audit [DBG] from='client.15384 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:53.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:52 vm07 bash[17031]: audit 2026-03-07T07:03:52.894853+0000 mon.vm07 (mon.0) 962 : audit [DBG] from='client.? 192.168.123.107:0/1668099927' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:53.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:52 vm07 bash[17031]: audit 2026-03-07T07:03:52.894853+0000 mon.vm07 (mon.0) 962 : audit [DBG] from='client.? 192.168.123.107:0/1668099927' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:54.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:53 vm09 bash[21220]: audit 2026-03-07T07:03:52.700355+0000 mgr.vm07.yrfcuj (mgr.14201) 512 : audit [DBG] from='client.15388 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:54.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:53 vm09 bash[21220]: audit 2026-03-07T07:03:52.700355+0000 mgr.vm07.yrfcuj (mgr.14201) 512 : audit [DBG] from='client.15388 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:54.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:53 vm07 bash[17031]: audit 2026-03-07T07:03:52.700355+0000 mgr.vm07.yrfcuj (mgr.14201) 512 : audit [DBG] from='client.15388 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:54.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:53 vm07 bash[17031]: audit 2026-03-07T07:03:52.700355+0000 mgr.vm07.yrfcuj (mgr.14201) 512 : audit [DBG] from='client.15388 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:55.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:54 vm09 bash[21220]: cluster 2026-03-07T07:03:53.703162+0000 mgr.vm07.yrfcuj (mgr.14201) 513 : cluster [DBG] pgmap v310: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:03:55.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:54 vm09 bash[21220]: cluster 2026-03-07T07:03:53.703162+0000 mgr.vm07.yrfcuj (mgr.14201) 513 : cluster [DBG] pgmap v310: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:03:55.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:54 vm07 bash[17031]: cluster 2026-03-07T07:03:53.703162+0000 mgr.vm07.yrfcuj (mgr.14201) 513 : cluster [DBG] pgmap v310: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:03:55.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:54 vm07 bash[17031]: cluster 2026-03-07T07:03:53.703162+0000 mgr.vm07.yrfcuj (mgr.14201) 513 : cluster [DBG] pgmap v310: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:03:56.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:55 vm09 bash[21220]: audit 2026-03-07T07:03:55.135405+0000 mon.vm07 (mon.0) 963 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:03:56.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:55 vm09 bash[21220]: audit 2026-03-07T07:03:55.135405+0000 mon.vm07 (mon.0) 963 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:03:56.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:55 vm09 bash[21220]: audit 2026-03-07T07:03:55.484275+0000 mon.vm07 (mon.0) 964 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:56.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:55 vm09 bash[21220]: audit 2026-03-07T07:03:55.484275+0000 mon.vm07 (mon.0) 964 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:56.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:55 vm09 bash[21220]: audit 2026-03-07T07:03:55.489927+0000 mon.vm07 (mon.0) 965 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:56.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:55 vm09 bash[21220]: audit 2026-03-07T07:03:55.489927+0000 mon.vm07 (mon.0) 965 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:56.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:55 vm09 bash[21220]: audit 2026-03-07T07:03:55.490918+0000 mon.vm07 (mon.0) 966 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:56.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:55 vm09 bash[21220]: audit 2026-03-07T07:03:55.490918+0000 mon.vm07 (mon.0) 966 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:56.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:55 vm09 bash[21220]: audit 2026-03-07T07:03:55.491461+0000 mon.vm07 (mon.0) 967 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:56.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:55 vm09 bash[21220]: audit 2026-03-07T07:03:55.491461+0000 mon.vm07 (mon.0) 967 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:56.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:55 vm09 bash[21220]: audit 2026-03-07T07:03:55.495447+0000 mon.vm07 (mon.0) 968 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:56.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:55 vm09 bash[21220]: audit 2026-03-07T07:03:55.495447+0000 mon.vm07 (mon.0) 968 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:56.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:55 vm09 bash[21220]: audit 2026-03-07T07:03:55.496797+0000 mon.vm07 (mon.0) 969 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:56.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:55 vm09 bash[21220]: audit 2026-03-07T07:03:55.496797+0000 mon.vm07 (mon.0) 969 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:56.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:55 vm07 bash[17031]: audit 2026-03-07T07:03:55.135405+0000 mon.vm07 (mon.0) 963 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:03:56.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:55 vm07 bash[17031]: audit 2026-03-07T07:03:55.135405+0000 mon.vm07 (mon.0) 963 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:03:56.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:55 vm07 bash[17031]: audit 2026-03-07T07:03:55.484275+0000 mon.vm07 (mon.0) 964 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:56.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:55 vm07 bash[17031]: audit 2026-03-07T07:03:55.484275+0000 mon.vm07 (mon.0) 964 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:56.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:55 vm07 bash[17031]: audit 2026-03-07T07:03:55.489927+0000 mon.vm07 (mon.0) 965 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:56.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:55 vm07 bash[17031]: audit 2026-03-07T07:03:55.489927+0000 mon.vm07 (mon.0) 965 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:56.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:55 vm07 bash[17031]: audit 2026-03-07T07:03:55.490918+0000 mon.vm07 (mon.0) 966 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:56.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:55 vm07 bash[17031]: audit 2026-03-07T07:03:55.490918+0000 mon.vm07 (mon.0) 966 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:56.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:55 vm07 bash[17031]: audit 2026-03-07T07:03:55.491461+0000 mon.vm07 (mon.0) 967 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:56.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:55 vm07 bash[17031]: audit 2026-03-07T07:03:55.491461+0000 mon.vm07 (mon.0) 967 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:56.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:55 vm07 bash[17031]: audit 2026-03-07T07:03:55.495447+0000 mon.vm07 (mon.0) 968 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:56.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:55 vm07 bash[17031]: audit 2026-03-07T07:03:55.495447+0000 mon.vm07 (mon.0) 968 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:56.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:55 vm07 bash[17031]: audit 2026-03-07T07:03:55.496797+0000 mon.vm07 (mon.0) 969 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:56.403 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:55 vm07 bash[17031]: audit 2026-03-07T07:03:55.496797+0000 mon.vm07 (mon.0) 969 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:57.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:57 vm09 bash[21220]: cluster 2026-03-07T07:03:55.703470+0000 mgr.vm07.yrfcuj (mgr.14201) 514 : cluster [DBG] pgmap v311: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:03:57.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:57 vm09 bash[21220]: cluster 2026-03-07T07:03:55.703470+0000 mgr.vm07.yrfcuj (mgr.14201) 514 : cluster [DBG] pgmap v311: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:03:57.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:57 vm07 bash[17031]: cluster 2026-03-07T07:03:55.703470+0000 mgr.vm07.yrfcuj (mgr.14201) 514 : cluster [DBG] pgmap v311: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:03:57.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:57 vm07 bash[17031]: cluster 2026-03-07T07:03:55.703470+0000 mgr.vm07.yrfcuj (mgr.14201) 514 : cluster [DBG] pgmap v311: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:03:58.075 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 stopped 2s ago 6m - - 2026-03-07T08:03:58.080 INFO:teuthology.orchestra.run.vm07.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-07T08:03:58.080 INFO:teuthology.orchestra.run.vm07.stderr: Dload Upload Total Spent Left Speed 2026-03-07T08:03:58.081 INFO:teuthology.orchestra.run.vm07.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 182k 0 --:--:-- --:--:-- --:--:-- 182k 2026-03-07T08:03:58.244 INFO:teuthology.orchestra.run.vm07.stdout:anonymousScheduled to start rgw.foo.vm07.lrxyjy on host 'vm07' 2026-03-07T08:03:58.442 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.lrxyjy to start 2026-03-07T08:03:58.608 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:03:58.608 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (30s) 3s ago 6m 92.6M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:03:58.608 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 stopped 3s ago 6m - - 2026-03-07T08:03:58.608 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (6m) 116s ago 6m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:03:58.608 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (6m) 116s ago 6m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:03:58.799 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: cluster 2026-03-07T07:03:57.703827+0000 mgr.vm07.yrfcuj (mgr.14201) 515 : cluster [DBG] pgmap v312: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: cluster 2026-03-07T07:03:57.703827+0000 mgr.vm07.yrfcuj (mgr.14201) 515 : cluster [DBG] pgmap v312: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.059473+0000 mgr.vm07.yrfcuj (mgr.14201) 516 : audit [DBG] from='client.15396 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.059473+0000 mgr.vm07.yrfcuj (mgr.14201) 516 : audit [DBG] from='client.15396 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.231330+0000 mgr.vm07.yrfcuj (mgr.14201) 517 : audit [DBG] from='client.15400 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "rgw.foo.vm07.lrxyjy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.231330+0000 mgr.vm07.yrfcuj (mgr.14201) 517 : audit [DBG] from='client.15400 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "rgw.foo.vm07.lrxyjy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: cephadm 2026-03-07T07:03:58.231675+0000 mgr.vm07.yrfcuj (mgr.14201) 518 : cephadm [INF] Schedule start daemon rgw.foo.vm07.lrxyjy 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: cephadm 2026-03-07T07:03:58.231675+0000 mgr.vm07.yrfcuj (mgr.14201) 518 : cephadm [INF] Schedule start daemon rgw.foo.vm07.lrxyjy 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.236969+0000 mon.vm07 (mon.0) 970 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.236969+0000 mon.vm07 (mon.0) 970 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.242230+0000 mon.vm07 (mon.0) 971 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.242230+0000 mon.vm07 (mon.0) 971 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.243181+0000 mon.vm07 (mon.0) 972 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.243181+0000 mon.vm07 (mon.0) 972 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.244113+0000 mon.vm07 (mon.0) 973 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.244113+0000 mon.vm07 (mon.0) 973 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.244514+0000 mon.vm07 (mon.0) 974 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.244514+0000 mon.vm07 (mon.0) 974 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.249752+0000 mon.vm07 (mon.0) 975 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.249752+0000 mon.vm07 (mon.0) 975 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.251387+0000 mon.vm07 (mon.0) 976 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.251387+0000 mon.vm07 (mon.0) 976 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.425003+0000 mgr.vm07.yrfcuj (mgr.14201) 519 : audit [DBG] from='client.15404 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.425003+0000 mgr.vm07.yrfcuj (mgr.14201) 519 : audit [DBG] from='client.15404 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.798840+0000 mon.vm09 (mon.1) 39 : audit [DBG] from='client.? 192.168.123.107:0/1800349684' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:59.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:03:59 vm09 bash[21220]: audit 2026-03-07T07:03:58.798840+0000 mon.vm09 (mon.1) 39 : audit [DBG] from='client.? 192.168.123.107:0/1800349684' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:59.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: cluster 2026-03-07T07:03:57.703827+0000 mgr.vm07.yrfcuj (mgr.14201) 515 : cluster [DBG] pgmap v312: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:03:59.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: cluster 2026-03-07T07:03:57.703827+0000 mgr.vm07.yrfcuj (mgr.14201) 515 : cluster [DBG] pgmap v312: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:03:59.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.059473+0000 mgr.vm07.yrfcuj (mgr.14201) 516 : audit [DBG] from='client.15396 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:59.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.059473+0000 mgr.vm07.yrfcuj (mgr.14201) 516 : audit [DBG] from='client.15396 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:59.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.231330+0000 mgr.vm07.yrfcuj (mgr.14201) 517 : audit [DBG] from='client.15400 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "rgw.foo.vm07.lrxyjy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:59.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.231330+0000 mgr.vm07.yrfcuj (mgr.14201) 517 : audit [DBG] from='client.15400 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "rgw.foo.vm07.lrxyjy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:59.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: cephadm 2026-03-07T07:03:58.231675+0000 mgr.vm07.yrfcuj (mgr.14201) 518 : cephadm [INF] Schedule start daemon rgw.foo.vm07.lrxyjy 2026-03-07T08:03:59.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: cephadm 2026-03-07T07:03:58.231675+0000 mgr.vm07.yrfcuj (mgr.14201) 518 : cephadm [INF] Schedule start daemon rgw.foo.vm07.lrxyjy 2026-03-07T08:03:59.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.236969+0000 mon.vm07 (mon.0) 970 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:59.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.236969+0000 mon.vm07 (mon.0) 970 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:59.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.242230+0000 mon.vm07 (mon.0) 971 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:59.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.242230+0000 mon.vm07 (mon.0) 971 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:59.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.243181+0000 mon.vm07 (mon.0) 972 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:59.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.243181+0000 mon.vm07 (mon.0) 972 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:03:59.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.244113+0000 mon.vm07 (mon.0) 973 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:59.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.244113+0000 mon.vm07 (mon.0) 973 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:03:59.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.244514+0000 mon.vm07 (mon.0) 974 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:59.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.244514+0000 mon.vm07 (mon.0) 974 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:03:59.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.249752+0000 mon.vm07 (mon.0) 975 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:59.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.249752+0000 mon.vm07 (mon.0) 975 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:03:59.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.251387+0000 mon.vm07 (mon.0) 976 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:59.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.251387+0000 mon.vm07 (mon.0) 976 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:03:59.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.425003+0000 mgr.vm07.yrfcuj (mgr.14201) 519 : audit [DBG] from='client.15404 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:59.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.425003+0000 mgr.vm07.yrfcuj (mgr.14201) 519 : audit [DBG] from='client.15404 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:03:59.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.798840+0000 mon.vm09 (mon.1) 39 : audit [DBG] from='client.? 192.168.123.107:0/1800349684' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:03:59.653 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:03:59 vm07 bash[17031]: audit 2026-03-07T07:03:58.798840+0000 mon.vm09 (mon.1) 39 : audit [DBG] from='client.? 192.168.123.107:0/1800349684' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:00 vm09 bash[21220]: audit 2026-03-07T07:03:58.605946+0000 mgr.vm07.yrfcuj (mgr.14201) 520 : audit [DBG] from='client.15408 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:00 vm09 bash[21220]: audit 2026-03-07T07:03:58.605946+0000 mgr.vm07.yrfcuj (mgr.14201) 520 : audit [DBG] from='client.15408 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:00.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:00 vm07 bash[17031]: audit 2026-03-07T07:03:58.605946+0000 mgr.vm07.yrfcuj (mgr.14201) 520 : audit [DBG] from='client.15408 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:00.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:00 vm07 bash[17031]: audit 2026-03-07T07:03:58.605946+0000 mgr.vm07.yrfcuj (mgr.14201) 520 : audit [DBG] from='client.15408 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:01.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:01 vm07 bash[17031]: cluster 2026-03-07T07:03:59.704159+0000 mgr.vm07.yrfcuj (mgr.14201) 521 : cluster [DBG] pgmap v313: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:04:01.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:01 vm07 bash[17031]: cluster 2026-03-07T07:03:59.704159+0000 mgr.vm07.yrfcuj (mgr.14201) 521 : cluster [DBG] pgmap v313: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:04:01.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:01 vm09 bash[21220]: cluster 2026-03-07T07:03:59.704159+0000 mgr.vm07.yrfcuj (mgr.14201) 521 : cluster [DBG] pgmap v313: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:04:01.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:01 vm09 bash[21220]: cluster 2026-03-07T07:03:59.704159+0000 mgr.vm07.yrfcuj (mgr.14201) 521 : cluster [DBG] pgmap v313: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:04:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:03 vm09 bash[21220]: cluster 2026-03-07T07:04:01.704519+0000 mgr.vm07.yrfcuj (mgr.14201) 522 : cluster [DBG] pgmap v314: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:03 vm09 bash[21220]: cluster 2026-03-07T07:04:01.704519+0000 mgr.vm07.yrfcuj (mgr.14201) 522 : cluster [DBG] pgmap v314: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:03.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:03 vm07 bash[17031]: cluster 2026-03-07T07:04:01.704519+0000 mgr.vm07.yrfcuj (mgr.14201) 522 : cluster [DBG] pgmap v314: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:03.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:03 vm07 bash[17031]: cluster 2026-03-07T07:04:01.704519+0000 mgr.vm07.yrfcuj (mgr.14201) 522 : cluster [DBG] pgmap v314: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:03.979 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.lrxyjy to start 2026-03-07T08:04:04.149 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:04:04.149 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (36s) 8s ago 6m 92.6M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:04:04.149 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 stopped 8s ago 6m - - 2026-03-07T08:04:04.149 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (6m) 2m ago 6m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:04:04.149 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (6m) 2m ago 6m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:04:04.339 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:04:05.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:05 vm09 bash[21220]: cluster 2026-03-07T07:04:03.704910+0000 mgr.vm07.yrfcuj (mgr.14201) 523 : cluster [DBG] pgmap v315: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:05.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:05 vm09 bash[21220]: cluster 2026-03-07T07:04:03.704910+0000 mgr.vm07.yrfcuj (mgr.14201) 523 : cluster [DBG] pgmap v315: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:05.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:05 vm09 bash[21220]: audit 2026-03-07T07:04:03.962246+0000 mgr.vm07.yrfcuj (mgr.14201) 524 : audit [DBG] from='client.15416 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:05.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:05 vm09 bash[21220]: audit 2026-03-07T07:04:03.962246+0000 mgr.vm07.yrfcuj (mgr.14201) 524 : audit [DBG] from='client.15416 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:05.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:05 vm09 bash[21220]: audit 2026-03-07T07:04:04.146592+0000 mgr.vm07.yrfcuj (mgr.14201) 525 : audit [DBG] from='client.15420 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:05.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:05 vm09 bash[21220]: audit 2026-03-07T07:04:04.146592+0000 mgr.vm07.yrfcuj (mgr.14201) 525 : audit [DBG] from='client.15420 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:05.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:05 vm09 bash[21220]: audit 2026-03-07T07:04:04.340192+0000 mon.vm07 (mon.0) 977 : audit [DBG] from='client.? 192.168.123.107:0/3136593933' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:05.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:05 vm09 bash[21220]: audit 2026-03-07T07:04:04.340192+0000 mon.vm07 (mon.0) 977 : audit [DBG] from='client.? 192.168.123.107:0/3136593933' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:05.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:05 vm07 bash[17031]: cluster 2026-03-07T07:04:03.704910+0000 mgr.vm07.yrfcuj (mgr.14201) 523 : cluster [DBG] pgmap v315: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:05.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:05 vm07 bash[17031]: cluster 2026-03-07T07:04:03.704910+0000 mgr.vm07.yrfcuj (mgr.14201) 523 : cluster [DBG] pgmap v315: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:05.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:05 vm07 bash[17031]: audit 2026-03-07T07:04:03.962246+0000 mgr.vm07.yrfcuj (mgr.14201) 524 : audit [DBG] from='client.15416 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:05.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:05 vm07 bash[17031]: audit 2026-03-07T07:04:03.962246+0000 mgr.vm07.yrfcuj (mgr.14201) 524 : audit [DBG] from='client.15416 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:05.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:05 vm07 bash[17031]: audit 2026-03-07T07:04:04.146592+0000 mgr.vm07.yrfcuj (mgr.14201) 525 : audit [DBG] from='client.15420 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:05.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:05 vm07 bash[17031]: audit 2026-03-07T07:04:04.146592+0000 mgr.vm07.yrfcuj (mgr.14201) 525 : audit [DBG] from='client.15420 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:05.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:05 vm07 bash[17031]: audit 2026-03-07T07:04:04.340192+0000 mon.vm07 (mon.0) 977 : audit [DBG] from='client.? 192.168.123.107:0/3136593933' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:05.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:05 vm07 bash[17031]: audit 2026-03-07T07:04:04.340192+0000 mon.vm07 (mon.0) 977 : audit [DBG] from='client.? 192.168.123.107:0/3136593933' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:07.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:07 vm07 bash[17031]: cluster 2026-03-07T07:04:05.705292+0000 mgr.vm07.yrfcuj (mgr.14201) 526 : cluster [DBG] pgmap v316: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:07.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:07 vm07 bash[17031]: cluster 2026-03-07T07:04:05.705292+0000 mgr.vm07.yrfcuj (mgr.14201) 526 : cluster [DBG] pgmap v316: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:07.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:07 vm07 bash[17031]: audit 2026-03-07T07:04:07.151739+0000 mon.vm07 (mon.0) 978 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:07.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:07 vm07 bash[17031]: audit 2026-03-07T07:04:07.151739+0000 mon.vm07 (mon.0) 978 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:07.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:07 vm07 bash[17031]: audit 2026-03-07T07:04:07.157181+0000 mon.vm07 (mon.0) 979 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:07.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:07 vm07 bash[17031]: audit 2026-03-07T07:04:07.157181+0000 mon.vm07 (mon.0) 979 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:07.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:07 vm07 bash[17031]: audit 2026-03-07T07:04:07.197573+0000 mon.vm07 (mon.0) 980 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:04:07.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:07 vm07 bash[17031]: audit 2026-03-07T07:04:07.197573+0000 mon.vm07 (mon.0) 980 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:04:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:07 vm09 bash[21220]: cluster 2026-03-07T07:04:05.705292+0000 mgr.vm07.yrfcuj (mgr.14201) 526 : cluster [DBG] pgmap v316: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:07 vm09 bash[21220]: cluster 2026-03-07T07:04:05.705292+0000 mgr.vm07.yrfcuj (mgr.14201) 526 : cluster [DBG] pgmap v316: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:07 vm09 bash[21220]: audit 2026-03-07T07:04:07.151739+0000 mon.vm07 (mon.0) 978 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:07 vm09 bash[21220]: audit 2026-03-07T07:04:07.151739+0000 mon.vm07 (mon.0) 978 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:07 vm09 bash[21220]: audit 2026-03-07T07:04:07.157181+0000 mon.vm07 (mon.0) 979 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:07 vm09 bash[21220]: audit 2026-03-07T07:04:07.157181+0000 mon.vm07 (mon.0) 979 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:07 vm09 bash[21220]: audit 2026-03-07T07:04:07.197573+0000 mon.vm07 (mon.0) 980 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:04:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:07 vm09 bash[21220]: audit 2026-03-07T07:04:07.197573+0000 mon.vm07 (mon.0) 980 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:04:09.536 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm07.lrxyjy to start 2026-03-07T08:04:09.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:09 vm09 bash[21220]: cluster 2026-03-07T07:04:07.705829+0000 mgr.vm07.yrfcuj (mgr.14201) 527 : cluster [DBG] pgmap v317: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 9.9 KiB/s rd, 170 B/s wr, 14 op/s 2026-03-07T08:04:09.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:09 vm09 bash[21220]: cluster 2026-03-07T07:04:07.705829+0000 mgr.vm07.yrfcuj (mgr.14201) 527 : cluster [DBG] pgmap v317: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 9.9 KiB/s rd, 170 B/s wr, 14 op/s 2026-03-07T08:04:09.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:09 vm07 bash[17031]: cluster 2026-03-07T07:04:07.705829+0000 mgr.vm07.yrfcuj (mgr.14201) 527 : cluster [DBG] pgmap v317: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 9.9 KiB/s rd, 170 B/s wr, 14 op/s 2026-03-07T08:04:09.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:09 vm07 bash[17031]: cluster 2026-03-07T07:04:07.705829+0000 mgr.vm07.yrfcuj (mgr.14201) 527 : cluster [DBG] pgmap v317: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 9.9 KiB/s rd, 170 B/s wr, 14 op/s 2026-03-07T08:04:09.696 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:04:09.696 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (41s) 14s ago 6m 92.6M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:04:09.696 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 stopped 14s ago 6m - - 2026-03-07T08:04:09.696 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (6m) 2m ago 6m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:04:09.696 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (6m) 2m ago 6m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:04:09.890 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:04:10.443 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:10 vm07 bash[17031]: audit 2026-03-07T07:04:09.521826+0000 mgr.vm07.yrfcuj (mgr.14201) 528 : audit [DBG] from='client.15440 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:10.443 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:10 vm07 bash[17031]: audit 2026-03-07T07:04:09.521826+0000 mgr.vm07.yrfcuj (mgr.14201) 528 : audit [DBG] from='client.15440 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:10.443 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:10 vm07 bash[17031]: audit 2026-03-07T07:04:09.890906+0000 mon.vm07 (mon.0) 981 : audit [DBG] from='client.? 192.168.123.107:0/1422647021' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:10.443 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:10 vm07 bash[17031]: audit 2026-03-07T07:04:09.890906+0000 mon.vm07 (mon.0) 981 : audit [DBG] from='client.? 192.168.123.107:0/1422647021' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:10.443 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:10 vm07 bash[17031]: audit 2026-03-07T07:04:10.135605+0000 mon.vm07 (mon.0) 982 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:04:10.443 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:10 vm07 bash[17031]: audit 2026-03-07T07:04:10.135605+0000 mon.vm07 (mon.0) 982 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:04:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:10 vm09 bash[21220]: audit 2026-03-07T07:04:09.521826+0000 mgr.vm07.yrfcuj (mgr.14201) 528 : audit [DBG] from='client.15440 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:10 vm09 bash[21220]: audit 2026-03-07T07:04:09.521826+0000 mgr.vm07.yrfcuj (mgr.14201) 528 : audit [DBG] from='client.15440 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:10 vm09 bash[21220]: audit 2026-03-07T07:04:09.890906+0000 mon.vm07 (mon.0) 981 : audit [DBG] from='client.? 192.168.123.107:0/1422647021' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:10 vm09 bash[21220]: audit 2026-03-07T07:04:09.890906+0000 mon.vm07 (mon.0) 981 : audit [DBG] from='client.? 192.168.123.107:0/1422647021' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:10 vm09 bash[21220]: audit 2026-03-07T07:04:10.135605+0000 mon.vm07 (mon.0) 982 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:04:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:10 vm09 bash[21220]: audit 2026-03-07T07:04:10.135605+0000 mon.vm07 (mon.0) 982 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:04:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:11 vm09 bash[21220]: audit 2026-03-07T07:04:09.693390+0000 mgr.vm07.yrfcuj (mgr.14201) 529 : audit [DBG] from='client.15444 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:11 vm09 bash[21220]: audit 2026-03-07T07:04:09.693390+0000 mgr.vm07.yrfcuj (mgr.14201) 529 : audit [DBG] from='client.15444 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:11 vm09 bash[21220]: cluster 2026-03-07T07:04:09.706229+0000 mgr.vm07.yrfcuj (mgr.14201) 530 : cluster [DBG] pgmap v318: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 20 KiB/s rd, 170 B/s wr, 30 op/s 2026-03-07T08:04:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:11 vm09 bash[21220]: cluster 2026-03-07T07:04:09.706229+0000 mgr.vm07.yrfcuj (mgr.14201) 530 : cluster [DBG] pgmap v318: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 20 KiB/s rd, 170 B/s wr, 30 op/s 2026-03-07T08:04:11.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:11 vm07 bash[17031]: audit 2026-03-07T07:04:09.693390+0000 mgr.vm07.yrfcuj (mgr.14201) 529 : audit [DBG] from='client.15444 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:11.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:11 vm07 bash[17031]: audit 2026-03-07T07:04:09.693390+0000 mgr.vm07.yrfcuj (mgr.14201) 529 : audit [DBG] from='client.15444 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:11.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:11 vm07 bash[17031]: cluster 2026-03-07T07:04:09.706229+0000 mgr.vm07.yrfcuj (mgr.14201) 530 : cluster [DBG] pgmap v318: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 20 KiB/s rd, 170 B/s wr, 30 op/s 2026-03-07T08:04:11.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:11 vm07 bash[17031]: cluster 2026-03-07T07:04:09.706229+0000 mgr.vm07.yrfcuj (mgr.14201) 530 : cluster [DBG] pgmap v318: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 20 KiB/s rd, 170 B/s wr, 30 op/s 2026-03-07T08:04:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:12 vm09 bash[21220]: cluster 2026-03-07T07:04:11.706634+0000 mgr.vm07.yrfcuj (mgr.14201) 531 : cluster [DBG] pgmap v319: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 38 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:04:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:12 vm09 bash[21220]: cluster 2026-03-07T07:04:11.706634+0000 mgr.vm07.yrfcuj (mgr.14201) 531 : cluster [DBG] pgmap v319: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 38 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:04:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:12 vm09 bash[21220]: audit 2026-03-07T07:04:11.955550+0000 mon.vm07 (mon.0) 983 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:12 vm09 bash[21220]: audit 2026-03-07T07:04:11.955550+0000 mon.vm07 (mon.0) 983 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:12 vm09 bash[21220]: audit 2026-03-07T07:04:11.961435+0000 mon.vm07 (mon.0) 984 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:12 vm09 bash[21220]: audit 2026-03-07T07:04:11.961435+0000 mon.vm07 (mon.0) 984 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:12 vm09 bash[21220]: audit 2026-03-07T07:04:12.272361+0000 mon.vm07 (mon.0) 985 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:04:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:12 vm09 bash[21220]: audit 2026-03-07T07:04:12.272361+0000 mon.vm07 (mon.0) 985 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:04:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:12 vm09 bash[21220]: audit 2026-03-07T07:04:12.272945+0000 mon.vm07 (mon.0) 986 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:04:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:12 vm09 bash[21220]: audit 2026-03-07T07:04:12.272945+0000 mon.vm07 (mon.0) 986 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:04:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:12 vm09 bash[21220]: audit 2026-03-07T07:04:12.277686+0000 mon.vm07 (mon.0) 987 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:12 vm09 bash[21220]: audit 2026-03-07T07:04:12.277686+0000 mon.vm07 (mon.0) 987 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:12 vm09 bash[21220]: audit 2026-03-07T07:04:12.279053+0000 mon.vm07 (mon.0) 988 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:04:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:12 vm09 bash[21220]: audit 2026-03-07T07:04:12.279053+0000 mon.vm07 (mon.0) 988 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:04:13.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:12 vm07 bash[17031]: cluster 2026-03-07T07:04:11.706634+0000 mgr.vm07.yrfcuj (mgr.14201) 531 : cluster [DBG] pgmap v319: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 38 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:04:13.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:12 vm07 bash[17031]: cluster 2026-03-07T07:04:11.706634+0000 mgr.vm07.yrfcuj (mgr.14201) 531 : cluster [DBG] pgmap v319: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 38 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:04:13.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:12 vm07 bash[17031]: audit 2026-03-07T07:04:11.955550+0000 mon.vm07 (mon.0) 983 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:13.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:12 vm07 bash[17031]: audit 2026-03-07T07:04:11.955550+0000 mon.vm07 (mon.0) 983 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:13.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:12 vm07 bash[17031]: audit 2026-03-07T07:04:11.961435+0000 mon.vm07 (mon.0) 984 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:13.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:12 vm07 bash[17031]: audit 2026-03-07T07:04:11.961435+0000 mon.vm07 (mon.0) 984 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:13.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:12 vm07 bash[17031]: audit 2026-03-07T07:04:12.272361+0000 mon.vm07 (mon.0) 985 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:04:13.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:12 vm07 bash[17031]: audit 2026-03-07T07:04:12.272361+0000 mon.vm07 (mon.0) 985 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:04:13.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:12 vm07 bash[17031]: audit 2026-03-07T07:04:12.272945+0000 mon.vm07 (mon.0) 986 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:04:13.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:12 vm07 bash[17031]: audit 2026-03-07T07:04:12.272945+0000 mon.vm07 (mon.0) 986 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:04:13.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:12 vm07 bash[17031]: audit 2026-03-07T07:04:12.277686+0000 mon.vm07 (mon.0) 987 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:13.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:12 vm07 bash[17031]: audit 2026-03-07T07:04:12.277686+0000 mon.vm07 (mon.0) 987 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:13.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:12 vm07 bash[17031]: audit 2026-03-07T07:04:12.279053+0000 mon.vm07 (mon.0) 988 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:04:13.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:12 vm07 bash[17031]: audit 2026-03-07T07:04:12.279053+0000 mon.vm07 (mon.0) 988 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:04:15.070 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (7s) 3s ago 6m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:04:15.260 INFO:teuthology.orchestra.run.vm07.stdout:Scheduled to stop rgw.foo.vm09.fgzozy on host 'vm09' 2026-03-07T08:04:15.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:14 vm09 bash[21220]: cluster 2026-03-07T07:04:13.707223+0000 mgr.vm07.yrfcuj (mgr.14201) 532 : cluster [DBG] pgmap v320: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 0 B/s wr, 82 op/s 2026-03-07T08:04:15.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:14 vm09 bash[21220]: cluster 2026-03-07T07:04:13.707223+0000 mgr.vm07.yrfcuj (mgr.14201) 532 : cluster [DBG] pgmap v320: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 0 B/s wr, 82 op/s 2026-03-07T08:04:15.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:14 vm07 bash[17031]: cluster 2026-03-07T07:04:13.707223+0000 mgr.vm07.yrfcuj (mgr.14201) 532 : cluster [DBG] pgmap v320: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 0 B/s wr, 82 op/s 2026-03-07T08:04:15.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:14 vm07 bash[17031]: cluster 2026-03-07T07:04:13.707223+0000 mgr.vm07.yrfcuj (mgr.14201) 532 : cluster [DBG] pgmap v320: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 0 B/s wr, 82 op/s 2026-03-07T08:04:15.459 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:04:15.620 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:04:15.621 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (47s) 3s ago 6m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:04:15.621 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (8s) 3s ago 6m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:04:15.621 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (6m) 2m ago 6m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:04:15.621 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (6m) 2m ago 6m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:04:15.806 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.055300+0000 mgr.vm07.yrfcuj (mgr.14201) 533 : audit [DBG] from='client.15452 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.055300+0000 mgr.vm07.yrfcuj (mgr.14201) 533 : audit [DBG] from='client.15452 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.244916+0000 mgr.vm07.yrfcuj (mgr.14201) 534 : audit [DBG] from='client.15456 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "rgw.foo.vm09.fgzozy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.244916+0000 mgr.vm07.yrfcuj (mgr.14201) 534 : audit [DBG] from='client.15456 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "rgw.foo.vm09.fgzozy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: cephadm 2026-03-07T07:04:15.245289+0000 mgr.vm07.yrfcuj (mgr.14201) 535 : cephadm [INF] Schedule stop daemon rgw.foo.vm09.fgzozy 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: cephadm 2026-03-07T07:04:15.245289+0000 mgr.vm07.yrfcuj (mgr.14201) 535 : cephadm [INF] Schedule stop daemon rgw.foo.vm09.fgzozy 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.254296+0000 mon.vm07 (mon.0) 989 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.254296+0000 mon.vm07 (mon.0) 989 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.259145+0000 mon.vm07 (mon.0) 990 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.259145+0000 mon.vm07 (mon.0) 990 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.259970+0000 mon.vm07 (mon.0) 991 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.259970+0000 mon.vm07 (mon.0) 991 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.261460+0000 mon.vm07 (mon.0) 992 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.261460+0000 mon.vm07 (mon.0) 992 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.261991+0000 mon.vm07 (mon.0) 993 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.261991+0000 mon.vm07 (mon.0) 993 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.265479+0000 mon.vm07 (mon.0) 994 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.265479+0000 mon.vm07 (mon.0) 994 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.266827+0000 mon.vm07 (mon.0) 995 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.266827+0000 mon.vm07 (mon.0) 995 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.443887+0000 mgr.vm07.yrfcuj (mgr.14201) 536 : audit [DBG] from='client.15460 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.443887+0000 mgr.vm07.yrfcuj (mgr.14201) 536 : audit [DBG] from='client.15460 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.806683+0000 mon.vm07 (mon.0) 996 : audit [DBG] from='client.? 192.168.123.107:0/2731269695' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:16 vm09 bash[21220]: audit 2026-03-07T07:04:15.806683+0000 mon.vm07 (mon.0) 996 : audit [DBG] from='client.? 192.168.123.107:0/2731269695' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.055300+0000 mgr.vm07.yrfcuj (mgr.14201) 533 : audit [DBG] from='client.15452 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.055300+0000 mgr.vm07.yrfcuj (mgr.14201) 533 : audit [DBG] from='client.15452 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.244916+0000 mgr.vm07.yrfcuj (mgr.14201) 534 : audit [DBG] from='client.15456 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "rgw.foo.vm09.fgzozy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.244916+0000 mgr.vm07.yrfcuj (mgr.14201) 534 : audit [DBG] from='client.15456 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "rgw.foo.vm09.fgzozy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: cephadm 2026-03-07T07:04:15.245289+0000 mgr.vm07.yrfcuj (mgr.14201) 535 : cephadm [INF] Schedule stop daemon rgw.foo.vm09.fgzozy 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: cephadm 2026-03-07T07:04:15.245289+0000 mgr.vm07.yrfcuj (mgr.14201) 535 : cephadm [INF] Schedule stop daemon rgw.foo.vm09.fgzozy 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.254296+0000 mon.vm07 (mon.0) 989 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.254296+0000 mon.vm07 (mon.0) 989 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.259145+0000 mon.vm07 (mon.0) 990 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.259145+0000 mon.vm07 (mon.0) 990 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.259970+0000 mon.vm07 (mon.0) 991 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.259970+0000 mon.vm07 (mon.0) 991 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.261460+0000 mon.vm07 (mon.0) 992 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.261460+0000 mon.vm07 (mon.0) 992 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.261991+0000 mon.vm07 (mon.0) 993 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.261991+0000 mon.vm07 (mon.0) 993 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.265479+0000 mon.vm07 (mon.0) 994 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.265479+0000 mon.vm07 (mon.0) 994 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.266827+0000 mon.vm07 (mon.0) 995 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.266827+0000 mon.vm07 (mon.0) 995 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.443887+0000 mgr.vm07.yrfcuj (mgr.14201) 536 : audit [DBG] from='client.15460 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.443887+0000 mgr.vm07.yrfcuj (mgr.14201) 536 : audit [DBG] from='client.15460 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.806683+0000 mon.vm07 (mon.0) 996 : audit [DBG] from='client.? 192.168.123.107:0/2731269695' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:16.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:16 vm07 bash[17031]: audit 2026-03-07T07:04:15.806683+0000 mon.vm07 (mon.0) 996 : audit [DBG] from='client.? 192.168.123.107:0/2731269695' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:17.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:17 vm09 bash[21220]: audit 2026-03-07T07:04:15.618776+0000 mgr.vm07.yrfcuj (mgr.14201) 537 : audit [DBG] from='client.15464 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:17.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:17 vm09 bash[21220]: audit 2026-03-07T07:04:15.618776+0000 mgr.vm07.yrfcuj (mgr.14201) 537 : audit [DBG] from='client.15464 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:17.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:17 vm09 bash[21220]: cluster 2026-03-07T07:04:15.707577+0000 mgr.vm07.yrfcuj (mgr.14201) 538 : cluster [DBG] pgmap v321: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 0 B/s wr, 82 op/s 2026-03-07T08:04:17.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:17 vm09 bash[21220]: cluster 2026-03-07T07:04:15.707577+0000 mgr.vm07.yrfcuj (mgr.14201) 538 : cluster [DBG] pgmap v321: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 0 B/s wr, 82 op/s 2026-03-07T08:04:17.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:17 vm07 bash[17031]: audit 2026-03-07T07:04:15.618776+0000 mgr.vm07.yrfcuj (mgr.14201) 537 : audit [DBG] from='client.15464 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:17.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:17 vm07 bash[17031]: audit 2026-03-07T07:04:15.618776+0000 mgr.vm07.yrfcuj (mgr.14201) 537 : audit [DBG] from='client.15464 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:17.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:17 vm07 bash[17031]: cluster 2026-03-07T07:04:15.707577+0000 mgr.vm07.yrfcuj (mgr.14201) 538 : cluster [DBG] pgmap v321: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 0 B/s wr, 82 op/s 2026-03-07T08:04:17.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:17 vm07 bash[17031]: cluster 2026-03-07T07:04:15.707577+0000 mgr.vm07.yrfcuj (mgr.14201) 538 : cluster [DBG] pgmap v321: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 0 B/s wr, 82 op/s 2026-03-07T08:04:19.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:19 vm09 bash[21220]: cluster 2026-03-07T07:04:17.708064+0000 mgr.vm07.yrfcuj (mgr.14201) 539 : cluster [DBG] pgmap v322: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 341 B/s wr, 82 op/s 2026-03-07T08:04:19.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:19 vm09 bash[21220]: cluster 2026-03-07T07:04:17.708064+0000 mgr.vm07.yrfcuj (mgr.14201) 539 : cluster [DBG] pgmap v322: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 341 B/s wr, 82 op/s 2026-03-07T08:04:19.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:19 vm07 bash[17031]: cluster 2026-03-07T07:04:17.708064+0000 mgr.vm07.yrfcuj (mgr.14201) 539 : cluster [DBG] pgmap v322: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 341 B/s wr, 82 op/s 2026-03-07T08:04:19.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:19 vm07 bash[17031]: cluster 2026-03-07T07:04:17.708064+0000 mgr.vm07.yrfcuj (mgr.14201) 539 : cluster [DBG] pgmap v322: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 341 B/s wr, 82 op/s 2026-03-07T08:04:20.985 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:04:21.136 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:04:21.136 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (53s) 9s ago 6m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:04:21.136 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (13s) 9s ago 6m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:04:21.136 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (6m) 2m ago 6m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:04:21.136 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (6m) 2m ago 6m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:04:21.330 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:04:21.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:21 vm09 bash[21220]: cluster 2026-03-07T07:04:19.708440+0000 mgr.vm07.yrfcuj (mgr.14201) 540 : cluster [DBG] pgmap v323: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 43 KiB/s rd, 341 B/s wr, 68 op/s 2026-03-07T08:04:21.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:21 vm09 bash[21220]: cluster 2026-03-07T07:04:19.708440+0000 mgr.vm07.yrfcuj (mgr.14201) 540 : cluster [DBG] pgmap v323: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 43 KiB/s rd, 341 B/s wr, 68 op/s 2026-03-07T08:04:21.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:21 vm07 bash[17031]: cluster 2026-03-07T07:04:19.708440+0000 mgr.vm07.yrfcuj (mgr.14201) 540 : cluster [DBG] pgmap v323: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 43 KiB/s rd, 341 B/s wr, 68 op/s 2026-03-07T08:04:21.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:21 vm07 bash[17031]: cluster 2026-03-07T07:04:19.708440+0000 mgr.vm07.yrfcuj (mgr.14201) 540 : cluster [DBG] pgmap v323: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 43 KiB/s rd, 341 B/s wr, 68 op/s 2026-03-07T08:04:22.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:22 vm09 bash[21220]: audit 2026-03-07T07:04:20.972092+0000 mgr.vm07.yrfcuj (mgr.14201) 541 : audit [DBG] from='client.15472 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:22.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:22 vm09 bash[21220]: audit 2026-03-07T07:04:20.972092+0000 mgr.vm07.yrfcuj (mgr.14201) 541 : audit [DBG] from='client.15472 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:22.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:22 vm09 bash[21220]: audit 2026-03-07T07:04:21.134315+0000 mgr.vm07.yrfcuj (mgr.14201) 542 : audit [DBG] from='client.15476 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:22.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:22 vm09 bash[21220]: audit 2026-03-07T07:04:21.134315+0000 mgr.vm07.yrfcuj (mgr.14201) 542 : audit [DBG] from='client.15476 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:22.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:22 vm09 bash[21220]: audit 2026-03-07T07:04:21.330895+0000 mon.vm07 (mon.0) 997 : audit [DBG] from='client.? 192.168.123.107:0/927668273' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:22.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:22 vm09 bash[21220]: audit 2026-03-07T07:04:21.330895+0000 mon.vm07 (mon.0) 997 : audit [DBG] from='client.? 192.168.123.107:0/927668273' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:22.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:22 vm07 bash[17031]: audit 2026-03-07T07:04:20.972092+0000 mgr.vm07.yrfcuj (mgr.14201) 541 : audit [DBG] from='client.15472 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:22.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:22 vm07 bash[17031]: audit 2026-03-07T07:04:20.972092+0000 mgr.vm07.yrfcuj (mgr.14201) 541 : audit [DBG] from='client.15472 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:22.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:22 vm07 bash[17031]: audit 2026-03-07T07:04:21.134315+0000 mgr.vm07.yrfcuj (mgr.14201) 542 : audit [DBG] from='client.15476 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:22.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:22 vm07 bash[17031]: audit 2026-03-07T07:04:21.134315+0000 mgr.vm07.yrfcuj (mgr.14201) 542 : audit [DBG] from='client.15476 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:22.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:22 vm07 bash[17031]: audit 2026-03-07T07:04:21.330895+0000 mon.vm07 (mon.0) 997 : audit [DBG] from='client.? 192.168.123.107:0/927668273' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:22.655 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:22 vm07 bash[17031]: audit 2026-03-07T07:04:21.330895+0000 mon.vm07 (mon.0) 997 : audit [DBG] from='client.? 192.168.123.107:0/927668273' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:23.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:23 vm09 bash[21220]: cluster 2026-03-07T07:04:21.708903+0000 mgr.vm07.yrfcuj (mgr.14201) 543 : cluster [DBG] pgmap v324: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 511 B/s wr, 53 op/s 2026-03-07T08:04:23.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:23 vm09 bash[21220]: cluster 2026-03-07T07:04:21.708903+0000 mgr.vm07.yrfcuj (mgr.14201) 543 : cluster [DBG] pgmap v324: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 511 B/s wr, 53 op/s 2026-03-07T08:04:23.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:23 vm07 bash[17031]: cluster 2026-03-07T07:04:21.708903+0000 mgr.vm07.yrfcuj (mgr.14201) 543 : cluster [DBG] pgmap v324: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 511 B/s wr, 53 op/s 2026-03-07T08:04:23.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:23 vm07 bash[17031]: cluster 2026-03-07T07:04:21.708903+0000 mgr.vm07.yrfcuj (mgr.14201) 543 : cluster [DBG] pgmap v324: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 511 B/s wr, 53 op/s 2026-03-07T08:04:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:25 vm09 bash[21220]: cluster 2026-03-07T07:04:23.709404+0000 mgr.vm07.yrfcuj (mgr.14201) 544 : cluster [DBG] pgmap v325: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 15 KiB/s rd, 511 B/s wr, 23 op/s 2026-03-07T08:04:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:25 vm09 bash[21220]: cluster 2026-03-07T07:04:23.709404+0000 mgr.vm07.yrfcuj (mgr.14201) 544 : cluster [DBG] pgmap v325: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 15 KiB/s rd, 511 B/s wr, 23 op/s 2026-03-07T08:04:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:25 vm09 bash[21220]: audit 2026-03-07T07:04:25.135679+0000 mon.vm07 (mon.0) 998 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:04:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:25 vm09 bash[21220]: audit 2026-03-07T07:04:25.135679+0000 mon.vm07 (mon.0) 998 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:04:25.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:25 vm07 bash[17031]: cluster 2026-03-07T07:04:23.709404+0000 mgr.vm07.yrfcuj (mgr.14201) 544 : cluster [DBG] pgmap v325: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 15 KiB/s rd, 511 B/s wr, 23 op/s 2026-03-07T08:04:25.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:25 vm07 bash[17031]: cluster 2026-03-07T07:04:23.709404+0000 mgr.vm07.yrfcuj (mgr.14201) 544 : cluster [DBG] pgmap v325: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 15 KiB/s rd, 511 B/s wr, 23 op/s 2026-03-07T08:04:25.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:25 vm07 bash[17031]: audit 2026-03-07T07:04:25.135679+0000 mon.vm07 (mon.0) 998 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:04:25.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:25 vm07 bash[17031]: audit 2026-03-07T07:04:25.135679+0000 mon.vm07 (mon.0) 998 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:04:26.517 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:04:26.671 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:04:26.671 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (58s) 14s ago 6m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:04:26.671 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (19s) 14s ago 6m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:04:26.671 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (6m) 2m ago 6m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:04:26.671 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (6m) 2m ago 6m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:04:26.856 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:04:27.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:27 vm09 bash[21220]: cluster 2026-03-07T07:04:25.709915+0000 mgr.vm07.yrfcuj (mgr.14201) 545 : cluster [DBG] pgmap v326: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:04:27.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:27 vm09 bash[21220]: cluster 2026-03-07T07:04:25.709915+0000 mgr.vm07.yrfcuj (mgr.14201) 545 : cluster [DBG] pgmap v326: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:04:27.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:27 vm09 bash[21220]: audit 2026-03-07T07:04:26.503433+0000 mgr.vm07.yrfcuj (mgr.14201) 546 : audit [DBG] from='client.15484 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:27.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:27 vm09 bash[21220]: audit 2026-03-07T07:04:26.503433+0000 mgr.vm07.yrfcuj (mgr.14201) 546 : audit [DBG] from='client.15484 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:27.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:27 vm09 bash[21220]: audit 2026-03-07T07:04:26.857706+0000 mon.vm07 (mon.0) 999 : audit [DBG] from='client.? 192.168.123.107:0/3031224402' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:27.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:27 vm09 bash[21220]: audit 2026-03-07T07:04:26.857706+0000 mon.vm07 (mon.0) 999 : audit [DBG] from='client.? 192.168.123.107:0/3031224402' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:27.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:27 vm07 bash[17031]: cluster 2026-03-07T07:04:25.709915+0000 mgr.vm07.yrfcuj (mgr.14201) 545 : cluster [DBG] pgmap v326: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:04:27.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:27 vm07 bash[17031]: cluster 2026-03-07T07:04:25.709915+0000 mgr.vm07.yrfcuj (mgr.14201) 545 : cluster [DBG] pgmap v326: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:04:27.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:27 vm07 bash[17031]: audit 2026-03-07T07:04:26.503433+0000 mgr.vm07.yrfcuj (mgr.14201) 546 : audit [DBG] from='client.15484 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:27.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:27 vm07 bash[17031]: audit 2026-03-07T07:04:26.503433+0000 mgr.vm07.yrfcuj (mgr.14201) 546 : audit [DBG] from='client.15484 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:27.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:27 vm07 bash[17031]: audit 2026-03-07T07:04:26.857706+0000 mon.vm07 (mon.0) 999 : audit [DBG] from='client.? 192.168.123.107:0/3031224402' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:27.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:27 vm07 bash[17031]: audit 2026-03-07T07:04:26.857706+0000 mon.vm07 (mon.0) 999 : audit [DBG] from='client.? 192.168.123.107:0/3031224402' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:28.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:28 vm09 bash[21220]: audit 2026-03-07T07:04:26.669345+0000 mgr.vm07.yrfcuj (mgr.14201) 547 : audit [DBG] from='client.15488 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:28.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:28 vm09 bash[21220]: audit 2026-03-07T07:04:26.669345+0000 mgr.vm07.yrfcuj (mgr.14201) 547 : audit [DBG] from='client.15488 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:28.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:28 vm07 bash[17031]: audit 2026-03-07T07:04:26.669345+0000 mgr.vm07.yrfcuj (mgr.14201) 547 : audit [DBG] from='client.15488 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:28.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:28 vm07 bash[17031]: audit 2026-03-07T07:04:26.669345+0000 mgr.vm07.yrfcuj (mgr.14201) 547 : audit [DBG] from='client.15488 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:29.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:29 vm09 bash[21220]: cluster 2026-03-07T07:04:27.710460+0000 mgr.vm07.yrfcuj (mgr.14201) 548 : cluster [DBG] pgmap v327: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:04:29.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:29 vm09 bash[21220]: cluster 2026-03-07T07:04:27.710460+0000 mgr.vm07.yrfcuj (mgr.14201) 548 : cluster [DBG] pgmap v327: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:04:29.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:29 vm07 bash[17031]: cluster 2026-03-07T07:04:27.710460+0000 mgr.vm07.yrfcuj (mgr.14201) 548 : cluster [DBG] pgmap v327: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:04:29.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:29 vm07 bash[17031]: cluster 2026-03-07T07:04:27.710460+0000 mgr.vm07.yrfcuj (mgr.14201) 548 : cluster [DBG] pgmap v327: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:04:31.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:31 vm09 bash[21220]: cluster 2026-03-07T07:04:29.710825+0000 mgr.vm07.yrfcuj (mgr.14201) 549 : cluster [DBG] pgmap v328: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:31.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:31 vm09 bash[21220]: cluster 2026-03-07T07:04:29.710825+0000 mgr.vm07.yrfcuj (mgr.14201) 549 : cluster [DBG] pgmap v328: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:31.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:31 vm09 bash[21220]: audit 2026-03-07T07:04:30.253888+0000 mon.vm07 (mon.0) 1000 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:31.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:31 vm09 bash[21220]: audit 2026-03-07T07:04:30.253888+0000 mon.vm07 (mon.0) 1000 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:31.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:31 vm09 bash[21220]: audit 2026-03-07T07:04:30.257931+0000 mon.vm07 (mon.0) 1001 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:31.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:31 vm09 bash[21220]: audit 2026-03-07T07:04:30.257931+0000 mon.vm07 (mon.0) 1001 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:31.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:31 vm09 bash[21220]: audit 2026-03-07T07:04:30.290434+0000 mon.vm07 (mon.0) 1002 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:04:31.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:31 vm09 bash[21220]: audit 2026-03-07T07:04:30.290434+0000 mon.vm07 (mon.0) 1002 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:04:31.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:31 vm09 bash[21220]: audit 2026-03-07T07:04:30.565742+0000 mon.vm07 (mon.0) 1003 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:31.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:31 vm09 bash[21220]: audit 2026-03-07T07:04:30.565742+0000 mon.vm07 (mon.0) 1003 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:31.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:31 vm09 bash[21220]: audit 2026-03-07T07:04:30.570388+0000 mon.vm07 (mon.0) 1004 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:31.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:31 vm09 bash[21220]: audit 2026-03-07T07:04:30.570388+0000 mon.vm07 (mon.0) 1004 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:31.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:31 vm07 bash[17031]: cluster 2026-03-07T07:04:29.710825+0000 mgr.vm07.yrfcuj (mgr.14201) 549 : cluster [DBG] pgmap v328: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:31.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:31 vm07 bash[17031]: cluster 2026-03-07T07:04:29.710825+0000 mgr.vm07.yrfcuj (mgr.14201) 549 : cluster [DBG] pgmap v328: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:31.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:31 vm07 bash[17031]: audit 2026-03-07T07:04:30.253888+0000 mon.vm07 (mon.0) 1000 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:31.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:31 vm07 bash[17031]: audit 2026-03-07T07:04:30.253888+0000 mon.vm07 (mon.0) 1000 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:31.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:31 vm07 bash[17031]: audit 2026-03-07T07:04:30.257931+0000 mon.vm07 (mon.0) 1001 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:31.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:31 vm07 bash[17031]: audit 2026-03-07T07:04:30.257931+0000 mon.vm07 (mon.0) 1001 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:31.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:31 vm07 bash[17031]: audit 2026-03-07T07:04:30.290434+0000 mon.vm07 (mon.0) 1002 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:04:31.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:31 vm07 bash[17031]: audit 2026-03-07T07:04:30.290434+0000 mon.vm07 (mon.0) 1002 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:04:31.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:31 vm07 bash[17031]: audit 2026-03-07T07:04:30.565742+0000 mon.vm07 (mon.0) 1003 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:31.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:31 vm07 bash[17031]: audit 2026-03-07T07:04:30.565742+0000 mon.vm07 (mon.0) 1003 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:31.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:31 vm07 bash[17031]: audit 2026-03-07T07:04:30.570388+0000 mon.vm07 (mon.0) 1004 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:31.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:31 vm07 bash[17031]: audit 2026-03-07T07:04:30.570388+0000 mon.vm07 (mon.0) 1004 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:32.028 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:04:32.191 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:04:32.191 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (64s) 20s ago 6m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:04:32.191 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (25s) 20s ago 6m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:04:32.191 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (6m) 2m ago 6m 101M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 6a1d83821630 2026-03-07T08:04:32.191 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (6m) 2m ago 6m 102M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:04:32.388 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:04:33.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:33 vm09 bash[21220]: cluster 2026-03-07T07:04:31.711229+0000 mgr.vm07.yrfcuj (mgr.14201) 550 : cluster [DBG] pgmap v329: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:04:33.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:33 vm09 bash[21220]: cluster 2026-03-07T07:04:31.711229+0000 mgr.vm07.yrfcuj (mgr.14201) 550 : cluster [DBG] pgmap v329: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:04:33.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:33 vm09 bash[21220]: audit 2026-03-07T07:04:32.014191+0000 mgr.vm07.yrfcuj (mgr.14201) 551 : audit [DBG] from='client.24969 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:33.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:33 vm09 bash[21220]: audit 2026-03-07T07:04:32.014191+0000 mgr.vm07.yrfcuj (mgr.14201) 551 : audit [DBG] from='client.24969 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:33.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:33 vm09 bash[21220]: audit 2026-03-07T07:04:32.189388+0000 mgr.vm07.yrfcuj (mgr.14201) 552 : audit [DBG] from='client.15500 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:33.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:33 vm09 bash[21220]: audit 2026-03-07T07:04:32.189388+0000 mgr.vm07.yrfcuj (mgr.14201) 552 : audit [DBG] from='client.15500 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:33.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:33 vm09 bash[21220]: audit 2026-03-07T07:04:32.389332+0000 mon.vm07 (mon.0) 1005 : audit [DBG] from='client.? 192.168.123.107:0/1722043346' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:33.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:33 vm09 bash[21220]: audit 2026-03-07T07:04:32.389332+0000 mon.vm07 (mon.0) 1005 : audit [DBG] from='client.? 192.168.123.107:0/1722043346' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:33.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:33 vm07 bash[17031]: cluster 2026-03-07T07:04:31.711229+0000 mgr.vm07.yrfcuj (mgr.14201) 550 : cluster [DBG] pgmap v329: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:04:33.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:33 vm07 bash[17031]: cluster 2026-03-07T07:04:31.711229+0000 mgr.vm07.yrfcuj (mgr.14201) 550 : cluster [DBG] pgmap v329: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:04:33.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:33 vm07 bash[17031]: audit 2026-03-07T07:04:32.014191+0000 mgr.vm07.yrfcuj (mgr.14201) 551 : audit [DBG] from='client.24969 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:33.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:33 vm07 bash[17031]: audit 2026-03-07T07:04:32.014191+0000 mgr.vm07.yrfcuj (mgr.14201) 551 : audit [DBG] from='client.24969 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:33.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:33 vm07 bash[17031]: audit 2026-03-07T07:04:32.189388+0000 mgr.vm07.yrfcuj (mgr.14201) 552 : audit [DBG] from='client.15500 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:33.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:33 vm07 bash[17031]: audit 2026-03-07T07:04:32.189388+0000 mgr.vm07.yrfcuj (mgr.14201) 552 : audit [DBG] from='client.15500 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:33.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:33 vm07 bash[17031]: audit 2026-03-07T07:04:32.389332+0000 mon.vm07 (mon.0) 1005 : audit [DBG] from='client.? 192.168.123.107:0/1722043346' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:33.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:33 vm07 bash[17031]: audit 2026-03-07T07:04:32.389332+0000 mon.vm07 (mon.0) 1005 : audit [DBG] from='client.? 192.168.123.107:0/1722043346' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:35.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:35 vm09 bash[21220]: cluster 2026-03-07T07:04:33.711590+0000 mgr.vm07.yrfcuj (mgr.14201) 553 : cluster [DBG] pgmap v330: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:35.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:35 vm09 bash[21220]: cluster 2026-03-07T07:04:33.711590+0000 mgr.vm07.yrfcuj (mgr.14201) 553 : cluster [DBG] pgmap v330: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:35.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:35 vm07 bash[17031]: cluster 2026-03-07T07:04:33.711590+0000 mgr.vm07.yrfcuj (mgr.14201) 553 : cluster [DBG] pgmap v330: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:35.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:35 vm07 bash[17031]: cluster 2026-03-07T07:04:33.711590+0000 mgr.vm07.yrfcuj (mgr.14201) 553 : cluster [DBG] pgmap v330: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:36 vm07 bash[17031]: cluster 2026-03-07T07:04:35.711990+0000 mgr.vm07.yrfcuj (mgr.14201) 554 : cluster [DBG] pgmap v331: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:36 vm07 bash[17031]: cluster 2026-03-07T07:04:35.711990+0000 mgr.vm07.yrfcuj (mgr.14201) 554 : cluster [DBG] pgmap v331: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:36 vm07 bash[17031]: audit 2026-03-07T07:04:35.803726+0000 mon.vm07 (mon.0) 1006 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:36 vm07 bash[17031]: audit 2026-03-07T07:04:35.803726+0000 mon.vm07 (mon.0) 1006 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:36 vm07 bash[17031]: audit 2026-03-07T07:04:35.808286+0000 mon.vm07 (mon.0) 1007 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:36 vm07 bash[17031]: audit 2026-03-07T07:04:35.808286+0000 mon.vm07 (mon.0) 1007 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:36 vm07 bash[17031]: audit 2026-03-07T07:04:35.809132+0000 mon.vm07 (mon.0) 1008 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:04:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:36 vm07 bash[17031]: audit 2026-03-07T07:04:35.809132+0000 mon.vm07 (mon.0) 1008 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:04:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:36 vm07 bash[17031]: audit 2026-03-07T07:04:35.809612+0000 mon.vm07 (mon.0) 1009 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:04:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:36 vm07 bash[17031]: audit 2026-03-07T07:04:35.809612+0000 mon.vm07 (mon.0) 1009 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:04:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:36 vm07 bash[17031]: cluster 2026-03-07T07:04:35.810822+0000 mgr.vm07.yrfcuj (mgr.14201) 555 : cluster [DBG] pgmap v332: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:36 vm07 bash[17031]: cluster 2026-03-07T07:04:35.810822+0000 mgr.vm07.yrfcuj (mgr.14201) 555 : cluster [DBG] pgmap v332: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:36 vm07 bash[17031]: audit 2026-03-07T07:04:35.813294+0000 mon.vm07 (mon.0) 1010 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:36 vm07 bash[17031]: audit 2026-03-07T07:04:35.813294+0000 mon.vm07 (mon.0) 1010 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:36 vm07 bash[17031]: audit 2026-03-07T07:04:35.814524+0000 mon.vm07 (mon.0) 1011 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:04:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:36 vm07 bash[17031]: audit 2026-03-07T07:04:35.814524+0000 mon.vm07 (mon.0) 1011 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:04:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:36 vm07 bash[17031]: cluster 2026-03-07T07:04:36.269649+0000 mon.vm07 (mon.0) 1012 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-07T08:04:36.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:36 vm07 bash[17031]: cluster 2026-03-07T07:04:36.269649+0000 mon.vm07 (mon.0) 1012 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-07T08:04:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:36 vm09 bash[21220]: cluster 2026-03-07T07:04:35.711990+0000 mgr.vm07.yrfcuj (mgr.14201) 554 : cluster [DBG] pgmap v331: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:36 vm09 bash[21220]: cluster 2026-03-07T07:04:35.711990+0000 mgr.vm07.yrfcuj (mgr.14201) 554 : cluster [DBG] pgmap v331: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:36 vm09 bash[21220]: audit 2026-03-07T07:04:35.803726+0000 mon.vm07 (mon.0) 1006 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:36 vm09 bash[21220]: audit 2026-03-07T07:04:35.803726+0000 mon.vm07 (mon.0) 1006 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:36 vm09 bash[21220]: audit 2026-03-07T07:04:35.808286+0000 mon.vm07 (mon.0) 1007 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:36 vm09 bash[21220]: audit 2026-03-07T07:04:35.808286+0000 mon.vm07 (mon.0) 1007 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:36 vm09 bash[21220]: audit 2026-03-07T07:04:35.809132+0000 mon.vm07 (mon.0) 1008 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:04:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:36 vm09 bash[21220]: audit 2026-03-07T07:04:35.809132+0000 mon.vm07 (mon.0) 1008 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:04:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:36 vm09 bash[21220]: audit 2026-03-07T07:04:35.809612+0000 mon.vm07 (mon.0) 1009 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:04:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:36 vm09 bash[21220]: audit 2026-03-07T07:04:35.809612+0000 mon.vm07 (mon.0) 1009 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:04:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:36 vm09 bash[21220]: cluster 2026-03-07T07:04:35.810822+0000 mgr.vm07.yrfcuj (mgr.14201) 555 : cluster [DBG] pgmap v332: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:36 vm09 bash[21220]: cluster 2026-03-07T07:04:35.810822+0000 mgr.vm07.yrfcuj (mgr.14201) 555 : cluster [DBG] pgmap v332: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:36 vm09 bash[21220]: audit 2026-03-07T07:04:35.813294+0000 mon.vm07 (mon.0) 1010 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:36 vm09 bash[21220]: audit 2026-03-07T07:04:35.813294+0000 mon.vm07 (mon.0) 1010 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:36 vm09 bash[21220]: audit 2026-03-07T07:04:35.814524+0000 mon.vm07 (mon.0) 1011 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:04:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:36 vm09 bash[21220]: audit 2026-03-07T07:04:35.814524+0000 mon.vm07 (mon.0) 1011 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:04:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:36 vm09 bash[21220]: cluster 2026-03-07T07:04:36.269649+0000 mon.vm07 (mon.0) 1012 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-07T08:04:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:36 vm09 bash[21220]: cluster 2026-03-07T07:04:36.269649+0000 mon.vm07 (mon.0) 1012 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-07T08:04:37.563 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:04:37.713 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:04:37.713 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (70s) 25s ago 6m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:04:37.713 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (30s) 25s ago 6m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:04:37.713 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 1s ago 7m - - 2026-03-07T08:04:37.713 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (6m) 1s ago 6m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:04:37.903 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:04:37.903 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:04:37.903 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:04:38.087 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:37 vm09 bash[21220]: audit 2026-03-07T07:04:37.548861+0000 mgr.vm07.yrfcuj (mgr.14201) 556 : audit [DBG] from='client.15508 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:38.087 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:37 vm09 bash[21220]: audit 2026-03-07T07:04:37.548861+0000 mgr.vm07.yrfcuj (mgr.14201) 556 : audit [DBG] from='client.15508 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:38.151 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:37 vm07 bash[17031]: audit 2026-03-07T07:04:37.548861+0000 mgr.vm07.yrfcuj (mgr.14201) 556 : audit [DBG] from='client.15508 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:38.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:37 vm07 bash[17031]: audit 2026-03-07T07:04:37.548861+0000 mgr.vm07.yrfcuj (mgr.14201) 556 : audit [DBG] from='client.15508 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:39.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:38 vm09 bash[21220]: audit 2026-03-07T07:04:37.712149+0000 mgr.vm07.yrfcuj (mgr.14201) 557 : audit [DBG] from='client.15512 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:39.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:38 vm09 bash[21220]: audit 2026-03-07T07:04:37.712149+0000 mgr.vm07.yrfcuj (mgr.14201) 557 : audit [DBG] from='client.15512 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:39.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:38 vm09 bash[21220]: cluster 2026-03-07T07:04:37.811323+0000 mgr.vm07.yrfcuj (mgr.14201) 558 : cluster [DBG] pgmap v333: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:39.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:38 vm09 bash[21220]: cluster 2026-03-07T07:04:37.811323+0000 mgr.vm07.yrfcuj (mgr.14201) 558 : cluster [DBG] pgmap v333: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:39.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:38 vm09 bash[21220]: audit 2026-03-07T07:04:37.903976+0000 mon.vm07 (mon.0) 1013 : audit [DBG] from='client.? 192.168.123.107:0/4121168859' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:39.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:38 vm09 bash[21220]: audit 2026-03-07T07:04:37.903976+0000 mon.vm07 (mon.0) 1013 : audit [DBG] from='client.? 192.168.123.107:0/4121168859' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:39.151 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:38 vm07 bash[17031]: audit 2026-03-07T07:04:37.712149+0000 mgr.vm07.yrfcuj (mgr.14201) 557 : audit [DBG] from='client.15512 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:39.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:38 vm07 bash[17031]: audit 2026-03-07T07:04:37.712149+0000 mgr.vm07.yrfcuj (mgr.14201) 557 : audit [DBG] from='client.15512 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:39.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:38 vm07 bash[17031]: cluster 2026-03-07T07:04:37.811323+0000 mgr.vm07.yrfcuj (mgr.14201) 558 : cluster [DBG] pgmap v333: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:39.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:38 vm07 bash[17031]: cluster 2026-03-07T07:04:37.811323+0000 mgr.vm07.yrfcuj (mgr.14201) 558 : cluster [DBG] pgmap v333: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:39.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:38 vm07 bash[17031]: audit 2026-03-07T07:04:37.903976+0000 mon.vm07 (mon.0) 1013 : audit [DBG] from='client.? 192.168.123.107:0/4121168859' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:39.152 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:38 vm07 bash[17031]: audit 2026-03-07T07:04:37.903976+0000 mon.vm07 (mon.0) 1013 : audit [DBG] from='client.? 192.168.123.107:0/4121168859' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:41.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:41 vm07 bash[17031]: cluster 2026-03-07T07:04:39.811765+0000 mgr.vm07.yrfcuj (mgr.14201) 559 : cluster [DBG] pgmap v334: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:41.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:41 vm07 bash[17031]: cluster 2026-03-07T07:04:39.811765+0000 mgr.vm07.yrfcuj (mgr.14201) 559 : cluster [DBG] pgmap v334: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:41.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:41 vm07 bash[17031]: audit 2026-03-07T07:04:40.139979+0000 mon.vm07 (mon.0) 1014 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:41.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:41 vm07 bash[17031]: audit 2026-03-07T07:04:40.139979+0000 mon.vm07 (mon.0) 1014 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:41.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:41 vm07 bash[17031]: audit 2026-03-07T07:04:40.140789+0000 mon.vm07 (mon.0) 1015 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:04:41.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:41 vm07 bash[17031]: audit 2026-03-07T07:04:40.140789+0000 mon.vm07 (mon.0) 1015 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:04:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:41 vm09 bash[21220]: cluster 2026-03-07T07:04:39.811765+0000 mgr.vm07.yrfcuj (mgr.14201) 559 : cluster [DBG] pgmap v334: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:41 vm09 bash[21220]: cluster 2026-03-07T07:04:39.811765+0000 mgr.vm07.yrfcuj (mgr.14201) 559 : cluster [DBG] pgmap v334: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:41 vm09 bash[21220]: audit 2026-03-07T07:04:40.139979+0000 mon.vm07 (mon.0) 1014 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:41 vm09 bash[21220]: audit 2026-03-07T07:04:40.139979+0000 mon.vm07 (mon.0) 1014 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:04:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:41 vm09 bash[21220]: audit 2026-03-07T07:04:40.140789+0000 mon.vm07 (mon.0) 1015 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:04:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:41 vm09 bash[21220]: audit 2026-03-07T07:04:40.140789+0000 mon.vm07 (mon.0) 1015 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:04:43.080 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:04:43.231 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:04:43.231 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (75s) 31s ago 7m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:04:43.231 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (36s) 31s ago 7m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:04:43.231 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 7s ago 7m - - 2026-03-07T08:04:43.231 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (7m) 7s ago 7m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:04:43.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:43 vm09 bash[21220]: cluster 2026-03-07T07:04:41.812144+0000 mgr.vm07.yrfcuj (mgr.14201) 560 : cluster [DBG] pgmap v335: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:43.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:43 vm09 bash[21220]: cluster 2026-03-07T07:04:41.812144+0000 mgr.vm07.yrfcuj (mgr.14201) 560 : cluster [DBG] pgmap v335: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:43.414 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:04:43.414 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:04:43.414 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:04:43.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:43 vm07 bash[17031]: cluster 2026-03-07T07:04:41.812144+0000 mgr.vm07.yrfcuj (mgr.14201) 560 : cluster [DBG] pgmap v335: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:43.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:43 vm07 bash[17031]: cluster 2026-03-07T07:04:41.812144+0000 mgr.vm07.yrfcuj (mgr.14201) 560 : cluster [DBG] pgmap v335: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:44.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:44 vm07 bash[17031]: audit 2026-03-07T07:04:43.066042+0000 mgr.vm07.yrfcuj (mgr.14201) 561 : audit [DBG] from='client.15520 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:44.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:44 vm07 bash[17031]: audit 2026-03-07T07:04:43.066042+0000 mgr.vm07.yrfcuj (mgr.14201) 561 : audit [DBG] from='client.15520 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:44.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:44 vm07 bash[17031]: audit 2026-03-07T07:04:43.230087+0000 mgr.vm07.yrfcuj (mgr.14201) 562 : audit [DBG] from='client.15524 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:44.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:44 vm07 bash[17031]: audit 2026-03-07T07:04:43.230087+0000 mgr.vm07.yrfcuj (mgr.14201) 562 : audit [DBG] from='client.15524 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:44.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:44 vm07 bash[17031]: audit 2026-03-07T07:04:43.413656+0000 mon.vm09 (mon.1) 40 : audit [DBG] from='client.? 192.168.123.107:0/3085887753' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:44.402 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:44 vm07 bash[17031]: audit 2026-03-07T07:04:43.413656+0000 mon.vm09 (mon.1) 40 : audit [DBG] from='client.? 192.168.123.107:0/3085887753' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:44 vm09 bash[21220]: audit 2026-03-07T07:04:43.066042+0000 mgr.vm07.yrfcuj (mgr.14201) 561 : audit [DBG] from='client.15520 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:44 vm09 bash[21220]: audit 2026-03-07T07:04:43.066042+0000 mgr.vm07.yrfcuj (mgr.14201) 561 : audit [DBG] from='client.15520 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:44 vm09 bash[21220]: audit 2026-03-07T07:04:43.230087+0000 mgr.vm07.yrfcuj (mgr.14201) 562 : audit [DBG] from='client.15524 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:44 vm09 bash[21220]: audit 2026-03-07T07:04:43.230087+0000 mgr.vm07.yrfcuj (mgr.14201) 562 : audit [DBG] from='client.15524 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:44 vm09 bash[21220]: audit 2026-03-07T07:04:43.413656+0000 mon.vm09 (mon.1) 40 : audit [DBG] from='client.? 192.168.123.107:0/3085887753' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:44 vm09 bash[21220]: audit 2026-03-07T07:04:43.413656+0000 mon.vm09 (mon.1) 40 : audit [DBG] from='client.? 192.168.123.107:0/3085887753' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:45 vm09 bash[21220]: cluster 2026-03-07T07:04:43.812575+0000 mgr.vm07.yrfcuj (mgr.14201) 563 : cluster [DBG] pgmap v336: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:45 vm09 bash[21220]: cluster 2026-03-07T07:04:43.812575+0000 mgr.vm07.yrfcuj (mgr.14201) 563 : cluster [DBG] pgmap v336: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:45.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:45 vm07 bash[17031]: cluster 2026-03-07T07:04:43.812575+0000 mgr.vm07.yrfcuj (mgr.14201) 563 : cluster [DBG] pgmap v336: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:45.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:45 vm07 bash[17031]: cluster 2026-03-07T07:04:43.812575+0000 mgr.vm07.yrfcuj (mgr.14201) 563 : cluster [DBG] pgmap v336: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:47.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:47 vm09 bash[21220]: cluster 2026-03-07T07:04:45.812959+0000 mgr.vm07.yrfcuj (mgr.14201) 564 : cluster [DBG] pgmap v337: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:47.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:47 vm09 bash[21220]: cluster 2026-03-07T07:04:45.812959+0000 mgr.vm07.yrfcuj (mgr.14201) 564 : cluster [DBG] pgmap v337: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:47.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:47 vm07 bash[17031]: cluster 2026-03-07T07:04:45.812959+0000 mgr.vm07.yrfcuj (mgr.14201) 564 : cluster [DBG] pgmap v337: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:47.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:47 vm07 bash[17031]: cluster 2026-03-07T07:04:45.812959+0000 mgr.vm07.yrfcuj (mgr.14201) 564 : cluster [DBG] pgmap v337: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 101 B/s rd, 202 B/s wr, 0 op/s 2026-03-07T08:04:48.590 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:04:48.740 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:04:48.740 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (81s) 36s ago 7m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:04:48.740 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (41s) 36s ago 7m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:04:48.740 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 12s ago 7m - - 2026-03-07T08:04:48.740 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (7m) 12s ago 7m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:04:48.926 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:04:48.926 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:04:48.926 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:04:49.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:49 vm07 bash[17031]: cluster 2026-03-07T07:04:47.813420+0000 mgr.vm07.yrfcuj (mgr.14201) 565 : cluster [DBG] pgmap v338: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:04:49.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:49 vm07 bash[17031]: cluster 2026-03-07T07:04:47.813420+0000 mgr.vm07.yrfcuj (mgr.14201) 565 : cluster [DBG] pgmap v338: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:04:49.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:49 vm07 bash[17031]: audit 2026-03-07T07:04:48.926507+0000 mon.vm07 (mon.0) 1016 : audit [DBG] from='client.? 192.168.123.107:0/2454643116' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:49.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:49 vm07 bash[17031]: audit 2026-03-07T07:04:48.926507+0000 mon.vm07 (mon.0) 1016 : audit [DBG] from='client.? 192.168.123.107:0/2454643116' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:49 vm09 bash[21220]: cluster 2026-03-07T07:04:47.813420+0000 mgr.vm07.yrfcuj (mgr.14201) 565 : cluster [DBG] pgmap v338: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:04:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:49 vm09 bash[21220]: cluster 2026-03-07T07:04:47.813420+0000 mgr.vm07.yrfcuj (mgr.14201) 565 : cluster [DBG] pgmap v338: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:04:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:49 vm09 bash[21220]: audit 2026-03-07T07:04:48.926507+0000 mon.vm07 (mon.0) 1016 : audit [DBG] from='client.? 192.168.123.107:0/2454643116' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:49 vm09 bash[21220]: audit 2026-03-07T07:04:48.926507+0000 mon.vm07 (mon.0) 1016 : audit [DBG] from='client.? 192.168.123.107:0/2454643116' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:50.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:50 vm07 bash[17031]: audit 2026-03-07T07:04:48.575224+0000 mgr.vm07.yrfcuj (mgr.14201) 566 : audit [DBG] from='client.15532 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:50.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:50 vm07 bash[17031]: audit 2026-03-07T07:04:48.575224+0000 mgr.vm07.yrfcuj (mgr.14201) 566 : audit [DBG] from='client.15532 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:50.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:50 vm07 bash[17031]: audit 2026-03-07T07:04:48.738814+0000 mgr.vm07.yrfcuj (mgr.14201) 567 : audit [DBG] from='client.15536 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:50.652 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:50 vm07 bash[17031]: audit 2026-03-07T07:04:48.738814+0000 mgr.vm07.yrfcuj (mgr.14201) 567 : audit [DBG] from='client.15536 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:50 vm09 bash[21220]: audit 2026-03-07T07:04:48.575224+0000 mgr.vm07.yrfcuj (mgr.14201) 566 : audit [DBG] from='client.15532 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:50 vm09 bash[21220]: audit 2026-03-07T07:04:48.575224+0000 mgr.vm07.yrfcuj (mgr.14201) 566 : audit [DBG] from='client.15532 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:50 vm09 bash[21220]: audit 2026-03-07T07:04:48.738814+0000 mgr.vm07.yrfcuj (mgr.14201) 567 : audit [DBG] from='client.15536 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:50 vm09 bash[21220]: audit 2026-03-07T07:04:48.738814+0000 mgr.vm07.yrfcuj (mgr.14201) 567 : audit [DBG] from='client.15536 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:51.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:51 vm07 bash[17031]: cluster 2026-03-07T07:04:49.813797+0000 mgr.vm07.yrfcuj (mgr.14201) 568 : cluster [DBG] pgmap v339: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:04:51.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:51 vm07 bash[17031]: cluster 2026-03-07T07:04:49.813797+0000 mgr.vm07.yrfcuj (mgr.14201) 568 : cluster [DBG] pgmap v339: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:04:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:51 vm09 bash[21220]: cluster 2026-03-07T07:04:49.813797+0000 mgr.vm07.yrfcuj (mgr.14201) 568 : cluster [DBG] pgmap v339: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:04:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:51 vm09 bash[21220]: cluster 2026-03-07T07:04:49.813797+0000 mgr.vm07.yrfcuj (mgr.14201) 568 : cluster [DBG] pgmap v339: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:04:53.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:53 vm07 bash[17031]: cluster 2026-03-07T07:04:51.814186+0000 mgr.vm07.yrfcuj (mgr.14201) 569 : cluster [DBG] pgmap v340: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:04:53.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:53 vm07 bash[17031]: cluster 2026-03-07T07:04:51.814186+0000 mgr.vm07.yrfcuj (mgr.14201) 569 : cluster [DBG] pgmap v340: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:04:53.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:53 vm09 bash[21220]: cluster 2026-03-07T07:04:51.814186+0000 mgr.vm07.yrfcuj (mgr.14201) 569 : cluster [DBG] pgmap v340: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:04:53.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:53 vm09 bash[21220]: cluster 2026-03-07T07:04:51.814186+0000 mgr.vm07.yrfcuj (mgr.14201) 569 : cluster [DBG] pgmap v340: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:04:54.101 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:04:54.255 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:04:54.255 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (86s) 42s ago 7m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:04:54.255 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (47s) 42s ago 7m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:04:54.255 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 18s ago 7m - - 2026-03-07T08:04:54.255 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (7m) 18s ago 7m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:04:54.455 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:04:54.455 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:04:54.455 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:04:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:55 vm09 bash[21220]: cluster 2026-03-07T07:04:53.814603+0000 mgr.vm07.yrfcuj (mgr.14201) 570 : cluster [DBG] pgmap v341: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:55 vm09 bash[21220]: cluster 2026-03-07T07:04:53.814603+0000 mgr.vm07.yrfcuj (mgr.14201) 570 : cluster [DBG] pgmap v341: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:55 vm09 bash[21220]: audit 2026-03-07T07:04:54.085410+0000 mgr.vm07.yrfcuj (mgr.14201) 571 : audit [DBG] from='client.15544 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:55 vm09 bash[21220]: audit 2026-03-07T07:04:54.085410+0000 mgr.vm07.yrfcuj (mgr.14201) 571 : audit [DBG] from='client.15544 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:55 vm09 bash[21220]: audit 2026-03-07T07:04:54.253982+0000 mgr.vm07.yrfcuj (mgr.14201) 572 : audit [DBG] from='client.15548 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:55 vm09 bash[21220]: audit 2026-03-07T07:04:54.253982+0000 mgr.vm07.yrfcuj (mgr.14201) 572 : audit [DBG] from='client.15548 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:55 vm09 bash[21220]: audit 2026-03-07T07:04:54.456139+0000 mon.vm07 (mon.0) 1017 : audit [DBG] from='client.? 192.168.123.107:0/4155783380' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:55 vm09 bash[21220]: audit 2026-03-07T07:04:54.456139+0000 mon.vm07 (mon.0) 1017 : audit [DBG] from='client.? 192.168.123.107:0/4155783380' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:55 vm09 bash[21220]: audit 2026-03-07T07:04:55.136342+0000 mon.vm07 (mon.0) 1018 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:04:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:55 vm09 bash[21220]: audit 2026-03-07T07:04:55.136342+0000 mon.vm07 (mon.0) 1018 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:04:55.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:55 vm07 bash[17031]: cluster 2026-03-07T07:04:53.814603+0000 mgr.vm07.yrfcuj (mgr.14201) 570 : cluster [DBG] pgmap v341: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:55.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:55 vm07 bash[17031]: cluster 2026-03-07T07:04:53.814603+0000 mgr.vm07.yrfcuj (mgr.14201) 570 : cluster [DBG] pgmap v341: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:55.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:55 vm07 bash[17031]: audit 2026-03-07T07:04:54.085410+0000 mgr.vm07.yrfcuj (mgr.14201) 571 : audit [DBG] from='client.15544 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:55.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:55 vm07 bash[17031]: audit 2026-03-07T07:04:54.085410+0000 mgr.vm07.yrfcuj (mgr.14201) 571 : audit [DBG] from='client.15544 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:55.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:55 vm07 bash[17031]: audit 2026-03-07T07:04:54.253982+0000 mgr.vm07.yrfcuj (mgr.14201) 572 : audit [DBG] from='client.15548 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:55.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:55 vm07 bash[17031]: audit 2026-03-07T07:04:54.253982+0000 mgr.vm07.yrfcuj (mgr.14201) 572 : audit [DBG] from='client.15548 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:04:55.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:55 vm07 bash[17031]: audit 2026-03-07T07:04:54.456139+0000 mon.vm07 (mon.0) 1017 : audit [DBG] from='client.? 192.168.123.107:0/4155783380' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:55.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:55 vm07 bash[17031]: audit 2026-03-07T07:04:54.456139+0000 mon.vm07 (mon.0) 1017 : audit [DBG] from='client.? 192.168.123.107:0/4155783380' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:04:55.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:55 vm07 bash[17031]: audit 2026-03-07T07:04:55.136342+0000 mon.vm07 (mon.0) 1018 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:04:55.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:55 vm07 bash[17031]: audit 2026-03-07T07:04:55.136342+0000 mon.vm07 (mon.0) 1018 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:04:57.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:57 vm09 bash[21220]: cluster 2026-03-07T07:04:55.815010+0000 mgr.vm07.yrfcuj (mgr.14201) 573 : cluster [DBG] pgmap v342: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:57.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:57 vm09 bash[21220]: cluster 2026-03-07T07:04:55.815010+0000 mgr.vm07.yrfcuj (mgr.14201) 573 : cluster [DBG] pgmap v342: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:57.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:57 vm07 bash[17031]: cluster 2026-03-07T07:04:55.815010+0000 mgr.vm07.yrfcuj (mgr.14201) 573 : cluster [DBG] pgmap v342: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:57.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:57 vm07 bash[17031]: cluster 2026-03-07T07:04:55.815010+0000 mgr.vm07.yrfcuj (mgr.14201) 573 : cluster [DBG] pgmap v342: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:59.639 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:04:59.804 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:04:59.804 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (92s) 47s ago 7m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:04:59.804 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (52s) 47s ago 7m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:04:59.804 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 24s ago 7m - - 2026-03-07T08:04:59.804 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (7m) 24s ago 7m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:04:59.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:59 vm09 bash[21220]: cluster 2026-03-07T07:04:57.815563+0000 mgr.vm07.yrfcuj (mgr.14201) 574 : cluster [DBG] pgmap v343: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:59.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:04:59 vm09 bash[21220]: cluster 2026-03-07T07:04:57.815563+0000 mgr.vm07.yrfcuj (mgr.14201) 574 : cluster [DBG] pgmap v343: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:59.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:59 vm07 bash[17031]: cluster 2026-03-07T07:04:57.815563+0000 mgr.vm07.yrfcuj (mgr.14201) 574 : cluster [DBG] pgmap v343: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:04:59.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:04:59 vm07 bash[17031]: cluster 2026-03-07T07:04:57.815563+0000 mgr.vm07.yrfcuj (mgr.14201) 574 : cluster [DBG] pgmap v343: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:00.008 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:05:00.008 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:05:00.008 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:05:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:00 vm09 bash[21220]: audit 2026-03-07T07:05:00.009232+0000 mon.vm07 (mon.0) 1019 : audit [DBG] from='client.? 192.168.123.107:0/3510321314' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:00 vm09 bash[21220]: audit 2026-03-07T07:05:00.009232+0000 mon.vm07 (mon.0) 1019 : audit [DBG] from='client.? 192.168.123.107:0/3510321314' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:00.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:00 vm07 bash[17031]: audit 2026-03-07T07:05:00.009232+0000 mon.vm07 (mon.0) 1019 : audit [DBG] from='client.? 192.168.123.107:0/3510321314' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:00.902 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:00 vm07 bash[17031]: audit 2026-03-07T07:05:00.009232+0000 mon.vm07 (mon.0) 1019 : audit [DBG] from='client.? 192.168.123.107:0/3510321314' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:01 vm09 bash[21220]: audit 2026-03-07T07:04:59.624472+0000 mgr.vm07.yrfcuj (mgr.14201) 575 : audit [DBG] from='client.25003 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:01 vm09 bash[21220]: audit 2026-03-07T07:04:59.624472+0000 mgr.vm07.yrfcuj (mgr.14201) 575 : audit [DBG] from='client.25003 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:01 vm09 bash[21220]: audit 2026-03-07T07:04:59.802565+0000 mgr.vm07.yrfcuj (mgr.14201) 576 : audit [DBG] from='client.15560 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:01 vm09 bash[21220]: audit 2026-03-07T07:04:59.802565+0000 mgr.vm07.yrfcuj (mgr.14201) 576 : audit [DBG] from='client.15560 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:01 vm09 bash[21220]: cluster 2026-03-07T07:04:59.816019+0000 mgr.vm07.yrfcuj (mgr.14201) 577 : cluster [DBG] pgmap v344: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:05:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:01 vm09 bash[21220]: cluster 2026-03-07T07:04:59.816019+0000 mgr.vm07.yrfcuj (mgr.14201) 577 : cluster [DBG] pgmap v344: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:05:01.866 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:01 vm07 bash[17031]: audit 2026-03-07T07:04:59.624472+0000 mgr.vm07.yrfcuj (mgr.14201) 575 : audit [DBG] from='client.25003 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:01.866 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:01 vm07 bash[17031]: audit 2026-03-07T07:04:59.624472+0000 mgr.vm07.yrfcuj (mgr.14201) 575 : audit [DBG] from='client.25003 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:01.866 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:01 vm07 bash[17031]: audit 2026-03-07T07:04:59.802565+0000 mgr.vm07.yrfcuj (mgr.14201) 576 : audit [DBG] from='client.15560 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:01.866 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:01 vm07 bash[17031]: audit 2026-03-07T07:04:59.802565+0000 mgr.vm07.yrfcuj (mgr.14201) 576 : audit [DBG] from='client.15560 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:01.866 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:01 vm07 bash[17031]: cluster 2026-03-07T07:04:59.816019+0000 mgr.vm07.yrfcuj (mgr.14201) 577 : cluster [DBG] pgmap v344: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:05:01.867 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:01 vm07 bash[17031]: cluster 2026-03-07T07:04:59.816019+0000 mgr.vm07.yrfcuj (mgr.14201) 577 : cluster [DBG] pgmap v344: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:05:03.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:03 vm09 bash[21220]: cluster 2026-03-07T07:05:01.816405+0000 mgr.vm07.yrfcuj (mgr.14201) 578 : cluster [DBG] pgmap v345: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:03.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:03 vm09 bash[21220]: cluster 2026-03-07T07:05:01.816405+0000 mgr.vm07.yrfcuj (mgr.14201) 578 : cluster [DBG] pgmap v345: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:03.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:03 vm07 bash[17031]: cluster 2026-03-07T07:05:01.816405+0000 mgr.vm07.yrfcuj (mgr.14201) 578 : cluster [DBG] pgmap v345: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:03.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:03 vm07 bash[17031]: cluster 2026-03-07T07:05:01.816405+0000 mgr.vm07.yrfcuj (mgr.14201) 578 : cluster [DBG] pgmap v345: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:05.200 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:05:05.364 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:05:05.364 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (97s) 53s ago 7m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:05:05.364 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (58s) 53s ago 7m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:05:05.364 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 29s ago 7m - - 2026-03-07T08:05:05.364 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (7m) 29s ago 7m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:05:05.569 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:05:05.569 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:05:05.569 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:05:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:05 vm09 bash[21220]: cluster 2026-03-07T07:05:03.816832+0000 mgr.vm07.yrfcuj (mgr.14201) 579 : cluster [DBG] pgmap v346: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:05 vm09 bash[21220]: cluster 2026-03-07T07:05:03.816832+0000 mgr.vm07.yrfcuj (mgr.14201) 579 : cluster [DBG] pgmap v346: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:05.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:05 vm07 bash[17031]: cluster 2026-03-07T07:05:03.816832+0000 mgr.vm07.yrfcuj (mgr.14201) 579 : cluster [DBG] pgmap v346: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:05.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:05 vm07 bash[17031]: cluster 2026-03-07T07:05:03.816832+0000 mgr.vm07.yrfcuj (mgr.14201) 579 : cluster [DBG] pgmap v346: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:06 vm09 bash[21220]: audit 2026-03-07T07:05:05.184840+0000 mgr.vm07.yrfcuj (mgr.14201) 580 : audit [DBG] from='client.15568 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:06 vm09 bash[21220]: audit 2026-03-07T07:05:05.184840+0000 mgr.vm07.yrfcuj (mgr.14201) 580 : audit [DBG] from='client.15568 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:06 vm09 bash[21220]: audit 2026-03-07T07:05:05.363139+0000 mgr.vm07.yrfcuj (mgr.14201) 581 : audit [DBG] from='client.15572 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:06 vm09 bash[21220]: audit 2026-03-07T07:05:05.363139+0000 mgr.vm07.yrfcuj (mgr.14201) 581 : audit [DBG] from='client.15572 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:06 vm09 bash[21220]: audit 2026-03-07T07:05:05.570612+0000 mon.vm07 (mon.0) 1020 : audit [DBG] from='client.? 192.168.123.107:0/671293389' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:06 vm09 bash[21220]: audit 2026-03-07T07:05:05.570612+0000 mon.vm07 (mon.0) 1020 : audit [DBG] from='client.? 192.168.123.107:0/671293389' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:06.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:06 vm07 bash[17031]: audit 2026-03-07T07:05:05.184840+0000 mgr.vm07.yrfcuj (mgr.14201) 580 : audit [DBG] from='client.15568 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:06.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:06 vm07 bash[17031]: audit 2026-03-07T07:05:05.184840+0000 mgr.vm07.yrfcuj (mgr.14201) 580 : audit [DBG] from='client.15568 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:06.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:06 vm07 bash[17031]: audit 2026-03-07T07:05:05.363139+0000 mgr.vm07.yrfcuj (mgr.14201) 581 : audit [DBG] from='client.15572 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:06.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:06 vm07 bash[17031]: audit 2026-03-07T07:05:05.363139+0000 mgr.vm07.yrfcuj (mgr.14201) 581 : audit [DBG] from='client.15572 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:06.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:06 vm07 bash[17031]: audit 2026-03-07T07:05:05.570612+0000 mon.vm07 (mon.0) 1020 : audit [DBG] from='client.? 192.168.123.107:0/671293389' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:06.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:06 vm07 bash[17031]: audit 2026-03-07T07:05:05.570612+0000 mon.vm07 (mon.0) 1020 : audit [DBG] from='client.? 192.168.123.107:0/671293389' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:07.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:07 vm09 bash[21220]: cluster 2026-03-07T07:05:05.817199+0000 mgr.vm07.yrfcuj (mgr.14201) 582 : cluster [DBG] pgmap v347: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:07.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:07 vm09 bash[21220]: cluster 2026-03-07T07:05:05.817199+0000 mgr.vm07.yrfcuj (mgr.14201) 582 : cluster [DBG] pgmap v347: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:07.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:07 vm07 bash[17031]: cluster 2026-03-07T07:05:05.817199+0000 mgr.vm07.yrfcuj (mgr.14201) 582 : cluster [DBG] pgmap v347: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:07.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:07 vm07 bash[17031]: cluster 2026-03-07T07:05:05.817199+0000 mgr.vm07.yrfcuj (mgr.14201) 582 : cluster [DBG] pgmap v347: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:09.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:09 vm09 bash[21220]: cluster 2026-03-07T07:05:07.817681+0000 mgr.vm07.yrfcuj (mgr.14201) 583 : cluster [DBG] pgmap v348: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:09.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:09 vm09 bash[21220]: cluster 2026-03-07T07:05:07.817681+0000 mgr.vm07.yrfcuj (mgr.14201) 583 : cluster [DBG] pgmap v348: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:09.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:09 vm07 bash[17031]: cluster 2026-03-07T07:05:07.817681+0000 mgr.vm07.yrfcuj (mgr.14201) 583 : cluster [DBG] pgmap v348: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:09.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:09 vm07 bash[17031]: cluster 2026-03-07T07:05:07.817681+0000 mgr.vm07.yrfcuj (mgr.14201) 583 : cluster [DBG] pgmap v348: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:10.745 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:05:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:10 vm09 bash[21220]: audit 2026-03-07T07:05:10.136477+0000 mon.vm07 (mon.0) 1021 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:05:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:10 vm09 bash[21220]: audit 2026-03-07T07:05:10.136477+0000 mon.vm07 (mon.0) 1021 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:05:10.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:10 vm07 bash[17031]: audit 2026-03-07T07:05:10.136477+0000 mon.vm07 (mon.0) 1021 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:05:10.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:10 vm07 bash[17031]: audit 2026-03-07T07:05:10.136477+0000 mon.vm07 (mon.0) 1021 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:05:10.907 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:05:10.907 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (103s) 58s ago 7m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:05:10.907 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (63s) 58s ago 7m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:05:10.907 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 35s ago 7m - - 2026-03-07T08:05:10.907 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (7m) 35s ago 7m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:05:11.101 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:05:11.101 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:05:11.101 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:05:11.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:11 vm09 bash[21220]: cluster 2026-03-07T07:05:09.818073+0000 mgr.vm07.yrfcuj (mgr.14201) 584 : cluster [DBG] pgmap v349: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:11.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:11 vm09 bash[21220]: cluster 2026-03-07T07:05:09.818073+0000 mgr.vm07.yrfcuj (mgr.14201) 584 : cluster [DBG] pgmap v349: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:11.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:11 vm09 bash[21220]: audit 2026-03-07T07:05:11.102379+0000 mon.vm07 (mon.0) 1022 : audit [DBG] from='client.? 192.168.123.107:0/4045913863' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:11.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:11 vm09 bash[21220]: audit 2026-03-07T07:05:11.102379+0000 mon.vm07 (mon.0) 1022 : audit [DBG] from='client.? 192.168.123.107:0/4045913863' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:11.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:11 vm07 bash[17031]: cluster 2026-03-07T07:05:09.818073+0000 mgr.vm07.yrfcuj (mgr.14201) 584 : cluster [DBG] pgmap v349: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:11.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:11 vm07 bash[17031]: cluster 2026-03-07T07:05:09.818073+0000 mgr.vm07.yrfcuj (mgr.14201) 584 : cluster [DBG] pgmap v349: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:11.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:11 vm07 bash[17031]: audit 2026-03-07T07:05:11.102379+0000 mon.vm07 (mon.0) 1022 : audit [DBG] from='client.? 192.168.123.107:0/4045913863' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:11.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:11 vm07 bash[17031]: audit 2026-03-07T07:05:11.102379+0000 mon.vm07 (mon.0) 1022 : audit [DBG] from='client.? 192.168.123.107:0/4045913863' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:12.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:12 vm09 bash[21220]: audit 2026-03-07T07:05:10.731810+0000 mgr.vm07.yrfcuj (mgr.14201) 585 : audit [DBG] from='client.15580 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:12.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:12 vm09 bash[21220]: audit 2026-03-07T07:05:10.731810+0000 mgr.vm07.yrfcuj (mgr.14201) 585 : audit [DBG] from='client.15580 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:12.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:12 vm09 bash[21220]: audit 2026-03-07T07:05:10.906096+0000 mgr.vm07.yrfcuj (mgr.14201) 586 : audit [DBG] from='client.15584 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:12.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:12 vm09 bash[21220]: audit 2026-03-07T07:05:10.906096+0000 mgr.vm07.yrfcuj (mgr.14201) 586 : audit [DBG] from='client.15584 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:12.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:12 vm07 bash[17031]: audit 2026-03-07T07:05:10.731810+0000 mgr.vm07.yrfcuj (mgr.14201) 585 : audit [DBG] from='client.15580 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:12.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:12 vm07 bash[17031]: audit 2026-03-07T07:05:10.731810+0000 mgr.vm07.yrfcuj (mgr.14201) 585 : audit [DBG] from='client.15580 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:12.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:12 vm07 bash[17031]: audit 2026-03-07T07:05:10.906096+0000 mgr.vm07.yrfcuj (mgr.14201) 586 : audit [DBG] from='client.15584 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:12.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:12 vm07 bash[17031]: audit 2026-03-07T07:05:10.906096+0000 mgr.vm07.yrfcuj (mgr.14201) 586 : audit [DBG] from='client.15584 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:13.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:13 vm09 bash[21220]: cluster 2026-03-07T07:05:11.818493+0000 mgr.vm07.yrfcuj (mgr.14201) 587 : cluster [DBG] pgmap v350: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:13.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:13 vm09 bash[21220]: cluster 2026-03-07T07:05:11.818493+0000 mgr.vm07.yrfcuj (mgr.14201) 587 : cluster [DBG] pgmap v350: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:13.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:13 vm07 bash[17031]: cluster 2026-03-07T07:05:11.818493+0000 mgr.vm07.yrfcuj (mgr.14201) 587 : cluster [DBG] pgmap v350: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:13.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:13 vm07 bash[17031]: cluster 2026-03-07T07:05:11.818493+0000 mgr.vm07.yrfcuj (mgr.14201) 587 : cluster [DBG] pgmap v350: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:15.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:15 vm09 bash[21220]: cluster 2026-03-07T07:05:13.819019+0000 mgr.vm07.yrfcuj (mgr.14201) 588 : cluster [DBG] pgmap v351: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:15.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:15 vm09 bash[21220]: cluster 2026-03-07T07:05:13.819019+0000 mgr.vm07.yrfcuj (mgr.14201) 588 : cluster [DBG] pgmap v351: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:15.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:15 vm07 bash[17031]: cluster 2026-03-07T07:05:13.819019+0000 mgr.vm07.yrfcuj (mgr.14201) 588 : cluster [DBG] pgmap v351: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:15.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:15 vm07 bash[17031]: cluster 2026-03-07T07:05:13.819019+0000 mgr.vm07.yrfcuj (mgr.14201) 588 : cluster [DBG] pgmap v351: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:16.267 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:05:16.411 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:05:16.411 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (108s) 64s ago 7m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:05:16.411 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (69s) 64s ago 7m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:05:16.411 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 40s ago 7m - - 2026-03-07T08:05:16.411 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (7m) 40s ago 7m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:05:16.593 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:05:16.593 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:05:16.593 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:05:16.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:16 vm09 bash[21220]: cluster 2026-03-07T07:05:15.819424+0000 mgr.vm07.yrfcuj (mgr.14201) 589 : cluster [DBG] pgmap v352: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:16.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:16 vm09 bash[21220]: cluster 2026-03-07T07:05:15.819424+0000 mgr.vm07.yrfcuj (mgr.14201) 589 : cluster [DBG] pgmap v352: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:16.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:16 vm09 bash[21220]: audit 2026-03-07T07:05:16.254966+0000 mgr.vm07.yrfcuj (mgr.14201) 590 : audit [DBG] from='client.15592 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:16.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:16 vm09 bash[21220]: audit 2026-03-07T07:05:16.254966+0000 mgr.vm07.yrfcuj (mgr.14201) 590 : audit [DBG] from='client.15592 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:16.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:16 vm09 bash[21220]: audit 2026-03-07T07:05:16.409848+0000 mgr.vm07.yrfcuj (mgr.14201) 591 : audit [DBG] from='client.15596 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:16.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:16 vm09 bash[21220]: audit 2026-03-07T07:05:16.409848+0000 mgr.vm07.yrfcuj (mgr.14201) 591 : audit [DBG] from='client.15596 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:16.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:16 vm07 bash[17031]: cluster 2026-03-07T07:05:15.819424+0000 mgr.vm07.yrfcuj (mgr.14201) 589 : cluster [DBG] pgmap v352: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:16.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:16 vm07 bash[17031]: cluster 2026-03-07T07:05:15.819424+0000 mgr.vm07.yrfcuj (mgr.14201) 589 : cluster [DBG] pgmap v352: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:16.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:16 vm07 bash[17031]: audit 2026-03-07T07:05:16.254966+0000 mgr.vm07.yrfcuj (mgr.14201) 590 : audit [DBG] from='client.15592 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:16.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:16 vm07 bash[17031]: audit 2026-03-07T07:05:16.254966+0000 mgr.vm07.yrfcuj (mgr.14201) 590 : audit [DBG] from='client.15592 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:16.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:16 vm07 bash[17031]: audit 2026-03-07T07:05:16.409848+0000 mgr.vm07.yrfcuj (mgr.14201) 591 : audit [DBG] from='client.15596 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:16.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:16 vm07 bash[17031]: audit 2026-03-07T07:05:16.409848+0000 mgr.vm07.yrfcuj (mgr.14201) 591 : audit [DBG] from='client.15596 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:17.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:17 vm09 bash[21220]: audit 2026-03-07T07:05:16.594483+0000 mon.vm07 (mon.0) 1023 : audit [DBG] from='client.? 192.168.123.107:0/3519760512' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:17.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:17 vm09 bash[21220]: audit 2026-03-07T07:05:16.594483+0000 mon.vm07 (mon.0) 1023 : audit [DBG] from='client.? 192.168.123.107:0/3519760512' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:17.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:17 vm07 bash[17031]: audit 2026-03-07T07:05:16.594483+0000 mon.vm07 (mon.0) 1023 : audit [DBG] from='client.? 192.168.123.107:0/3519760512' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:17.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:17 vm07 bash[17031]: audit 2026-03-07T07:05:16.594483+0000 mon.vm07 (mon.0) 1023 : audit [DBG] from='client.? 192.168.123.107:0/3519760512' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:18 vm09 bash[21220]: cluster 2026-03-07T07:05:17.819940+0000 mgr.vm07.yrfcuj (mgr.14201) 592 : cluster [DBG] pgmap v353: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:18 vm09 bash[21220]: cluster 2026-03-07T07:05:17.819940+0000 mgr.vm07.yrfcuj (mgr.14201) 592 : cluster [DBG] pgmap v353: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:18.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:18 vm07 bash[17031]: cluster 2026-03-07T07:05:17.819940+0000 mgr.vm07.yrfcuj (mgr.14201) 592 : cluster [DBG] pgmap v353: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:18.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:18 vm07 bash[17031]: cluster 2026-03-07T07:05:17.819940+0000 mgr.vm07.yrfcuj (mgr.14201) 592 : cluster [DBG] pgmap v353: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:21.151 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:20 vm07 bash[17031]: cluster 2026-03-07T07:05:19.820342+0000 mgr.vm07.yrfcuj (mgr.14201) 593 : cluster [DBG] pgmap v354: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:21.151 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:20 vm07 bash[17031]: cluster 2026-03-07T07:05:19.820342+0000 mgr.vm07.yrfcuj (mgr.14201) 593 : cluster [DBG] pgmap v354: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:21.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:20 vm09 bash[21220]: cluster 2026-03-07T07:05:19.820342+0000 mgr.vm07.yrfcuj (mgr.14201) 593 : cluster [DBG] pgmap v354: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:21.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:20 vm09 bash[21220]: cluster 2026-03-07T07:05:19.820342+0000 mgr.vm07.yrfcuj (mgr.14201) 593 : cluster [DBG] pgmap v354: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:21.775 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:05:21.940 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:05:21.940 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (114s) 70s ago 7m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:05:21.940 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (74s) 70s ago 7m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:05:21.940 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 46s ago 7m - - 2026-03-07T08:05:21.940 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (7m) 46s ago 7m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:05:22.141 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:05:22.141 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:05:22.141 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:05:23.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:23 vm09 bash[21220]: audit 2026-03-07T07:05:21.761944+0000 mgr.vm07.yrfcuj (mgr.14201) 594 : audit [DBG] from='client.15604 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:23.371 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:23 vm09 bash[21220]: audit 2026-03-07T07:05:21.761944+0000 mgr.vm07.yrfcuj (mgr.14201) 594 : audit [DBG] from='client.15604 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:23.371 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:23 vm09 bash[21220]: cluster 2026-03-07T07:05:21.820644+0000 mgr.vm07.yrfcuj (mgr.14201) 595 : cluster [DBG] pgmap v355: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:23.371 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:23 vm09 bash[21220]: cluster 2026-03-07T07:05:21.820644+0000 mgr.vm07.yrfcuj (mgr.14201) 595 : cluster [DBG] pgmap v355: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:23.371 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:23 vm09 bash[21220]: audit 2026-03-07T07:05:21.938625+0000 mgr.vm07.yrfcuj (mgr.14201) 596 : audit [DBG] from='client.15608 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:23.371 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:23 vm09 bash[21220]: audit 2026-03-07T07:05:21.938625+0000 mgr.vm07.yrfcuj (mgr.14201) 596 : audit [DBG] from='client.15608 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:23.371 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:23 vm09 bash[21220]: audit 2026-03-07T07:05:22.143079+0000 mon.vm07 (mon.0) 1024 : audit [DBG] from='client.? 192.168.123.107:0/1242933634' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:23.371 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:23 vm09 bash[21220]: audit 2026-03-07T07:05:22.143079+0000 mon.vm07 (mon.0) 1024 : audit [DBG] from='client.? 192.168.123.107:0/1242933634' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:23.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:23 vm07 bash[17031]: audit 2026-03-07T07:05:21.761944+0000 mgr.vm07.yrfcuj (mgr.14201) 594 : audit [DBG] from='client.15604 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:23.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:23 vm07 bash[17031]: audit 2026-03-07T07:05:21.761944+0000 mgr.vm07.yrfcuj (mgr.14201) 594 : audit [DBG] from='client.15604 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:23.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:23 vm07 bash[17031]: cluster 2026-03-07T07:05:21.820644+0000 mgr.vm07.yrfcuj (mgr.14201) 595 : cluster [DBG] pgmap v355: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:23.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:23 vm07 bash[17031]: cluster 2026-03-07T07:05:21.820644+0000 mgr.vm07.yrfcuj (mgr.14201) 595 : cluster [DBG] pgmap v355: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:23.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:23 vm07 bash[17031]: audit 2026-03-07T07:05:21.938625+0000 mgr.vm07.yrfcuj (mgr.14201) 596 : audit [DBG] from='client.15608 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:23.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:23 vm07 bash[17031]: audit 2026-03-07T07:05:21.938625+0000 mgr.vm07.yrfcuj (mgr.14201) 596 : audit [DBG] from='client.15608 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:23.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:23 vm07 bash[17031]: audit 2026-03-07T07:05:22.143079+0000 mon.vm07 (mon.0) 1024 : audit [DBG] from='client.? 192.168.123.107:0/1242933634' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:23.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:23 vm07 bash[17031]: audit 2026-03-07T07:05:22.143079+0000 mon.vm07 (mon.0) 1024 : audit [DBG] from='client.? 192.168.123.107:0/1242933634' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:25 vm09 bash[21220]: cluster 2026-03-07T07:05:23.821011+0000 mgr.vm07.yrfcuj (mgr.14201) 597 : cluster [DBG] pgmap v356: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:25 vm09 bash[21220]: cluster 2026-03-07T07:05:23.821011+0000 mgr.vm07.yrfcuj (mgr.14201) 597 : cluster [DBG] pgmap v356: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:25.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:25 vm07 bash[17031]: cluster 2026-03-07T07:05:23.821011+0000 mgr.vm07.yrfcuj (mgr.14201) 597 : cluster [DBG] pgmap v356: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:25.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:25 vm07 bash[17031]: cluster 2026-03-07T07:05:23.821011+0000 mgr.vm07.yrfcuj (mgr.14201) 597 : cluster [DBG] pgmap v356: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:26.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:26 vm09 bash[21220]: audit 2026-03-07T07:05:25.136614+0000 mon.vm07 (mon.0) 1025 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:05:26.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:26 vm09 bash[21220]: audit 2026-03-07T07:05:25.136614+0000 mon.vm07 (mon.0) 1025 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:05:26.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:26 vm07 bash[17031]: audit 2026-03-07T07:05:25.136614+0000 mon.vm07 (mon.0) 1025 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:05:26.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:26 vm07 bash[17031]: audit 2026-03-07T07:05:25.136614+0000 mon.vm07 (mon.0) 1025 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:05:27.341 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:05:27.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:27 vm09 bash[21220]: cluster 2026-03-07T07:05:25.821366+0000 mgr.vm07.yrfcuj (mgr.14201) 598 : cluster [DBG] pgmap v357: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:27.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:27 vm09 bash[21220]: cluster 2026-03-07T07:05:25.821366+0000 mgr.vm07.yrfcuj (mgr.14201) 598 : cluster [DBG] pgmap v357: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:27.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:27 vm07 bash[17031]: cluster 2026-03-07T07:05:25.821366+0000 mgr.vm07.yrfcuj (mgr.14201) 598 : cluster [DBG] pgmap v357: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:27.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:27 vm07 bash[17031]: cluster 2026-03-07T07:05:25.821366+0000 mgr.vm07.yrfcuj (mgr.14201) 598 : cluster [DBG] pgmap v357: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:27.491 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:05:27.491 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (119s) 75s ago 7m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:05:27.491 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (80s) 75s ago 7m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:05:27.491 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 51s ago 7m - - 2026-03-07T08:05:27.491 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (7m) 51s ago 7m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:05:27.677 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:05:27.677 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:05:27.677 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:05:28.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:28 vm09 bash[21220]: audit 2026-03-07T07:05:27.328551+0000 mgr.vm07.yrfcuj (mgr.14201) 599 : audit [DBG] from='client.15616 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:28.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:28 vm09 bash[21220]: audit 2026-03-07T07:05:27.328551+0000 mgr.vm07.yrfcuj (mgr.14201) 599 : audit [DBG] from='client.15616 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:28.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:28 vm09 bash[21220]: audit 2026-03-07T07:05:27.490069+0000 mgr.vm07.yrfcuj (mgr.14201) 600 : audit [DBG] from='client.15620 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:28.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:28 vm09 bash[21220]: audit 2026-03-07T07:05:27.490069+0000 mgr.vm07.yrfcuj (mgr.14201) 600 : audit [DBG] from='client.15620 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:28.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:28 vm09 bash[21220]: audit 2026-03-07T07:05:27.678507+0000 mon.vm07 (mon.0) 1026 : audit [DBG] from='client.? 192.168.123.107:0/2164276706' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:28.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:28 vm09 bash[21220]: audit 2026-03-07T07:05:27.678507+0000 mon.vm07 (mon.0) 1026 : audit [DBG] from='client.? 192.168.123.107:0/2164276706' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:28.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:28 vm07 bash[17031]: audit 2026-03-07T07:05:27.328551+0000 mgr.vm07.yrfcuj (mgr.14201) 599 : audit [DBG] from='client.15616 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:28.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:28 vm07 bash[17031]: audit 2026-03-07T07:05:27.328551+0000 mgr.vm07.yrfcuj (mgr.14201) 599 : audit [DBG] from='client.15616 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:28.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:28 vm07 bash[17031]: audit 2026-03-07T07:05:27.490069+0000 mgr.vm07.yrfcuj (mgr.14201) 600 : audit [DBG] from='client.15620 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:28.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:28 vm07 bash[17031]: audit 2026-03-07T07:05:27.490069+0000 mgr.vm07.yrfcuj (mgr.14201) 600 : audit [DBG] from='client.15620 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:28.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:28 vm07 bash[17031]: audit 2026-03-07T07:05:27.678507+0000 mon.vm07 (mon.0) 1026 : audit [DBG] from='client.? 192.168.123.107:0/2164276706' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:28.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:28 vm07 bash[17031]: audit 2026-03-07T07:05:27.678507+0000 mon.vm07 (mon.0) 1026 : audit [DBG] from='client.? 192.168.123.107:0/2164276706' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:29.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:29 vm09 bash[21220]: cluster 2026-03-07T07:05:27.821924+0000 mgr.vm07.yrfcuj (mgr.14201) 601 : cluster [DBG] pgmap v358: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:29.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:29 vm09 bash[21220]: cluster 2026-03-07T07:05:27.821924+0000 mgr.vm07.yrfcuj (mgr.14201) 601 : cluster [DBG] pgmap v358: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:29.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:29 vm07 bash[17031]: cluster 2026-03-07T07:05:27.821924+0000 mgr.vm07.yrfcuj (mgr.14201) 601 : cluster [DBG] pgmap v358: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:29.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:29 vm07 bash[17031]: cluster 2026-03-07T07:05:27.821924+0000 mgr.vm07.yrfcuj (mgr.14201) 601 : cluster [DBG] pgmap v358: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:31.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:31 vm09 bash[21220]: cluster 2026-03-07T07:05:29.822315+0000 mgr.vm07.yrfcuj (mgr.14201) 602 : cluster [DBG] pgmap v359: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:05:31.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:31 vm09 bash[21220]: cluster 2026-03-07T07:05:29.822315+0000 mgr.vm07.yrfcuj (mgr.14201) 602 : cluster [DBG] pgmap v359: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:05:31.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:31 vm07 bash[17031]: cluster 2026-03-07T07:05:29.822315+0000 mgr.vm07.yrfcuj (mgr.14201) 602 : cluster [DBG] pgmap v359: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:05:31.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:31 vm07 bash[17031]: cluster 2026-03-07T07:05:29.822315+0000 mgr.vm07.yrfcuj (mgr.14201) 602 : cluster [DBG] pgmap v359: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:05:32.854 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:05:33.010 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:05:33.010 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (2m) 81s ago 7m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:05:33.010 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (85s) 81s ago 7m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:05:33.010 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 57s ago 7m - - 2026-03-07T08:05:33.010 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (7m) 57s ago 7m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:05:33.207 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:05:33.207 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:05:33.207 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:05:33.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:33 vm09 bash[21220]: cluster 2026-03-07T07:05:31.822703+0000 mgr.vm07.yrfcuj (mgr.14201) 603 : cluster [DBG] pgmap v360: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:33.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:33 vm09 bash[21220]: cluster 2026-03-07T07:05:31.822703+0000 mgr.vm07.yrfcuj (mgr.14201) 603 : cluster [DBG] pgmap v360: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:33.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:33 vm07 bash[17031]: cluster 2026-03-07T07:05:31.822703+0000 mgr.vm07.yrfcuj (mgr.14201) 603 : cluster [DBG] pgmap v360: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:33.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:33 vm07 bash[17031]: cluster 2026-03-07T07:05:31.822703+0000 mgr.vm07.yrfcuj (mgr.14201) 603 : cluster [DBG] pgmap v360: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:34.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:34 vm07 bash[17031]: audit 2026-03-07T07:05:32.840594+0000 mgr.vm07.yrfcuj (mgr.14201) 604 : audit [DBG] from='client.15628 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:34.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:34 vm07 bash[17031]: audit 2026-03-07T07:05:32.840594+0000 mgr.vm07.yrfcuj (mgr.14201) 604 : audit [DBG] from='client.15628 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:34.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:34 vm07 bash[17031]: audit 2026-03-07T07:05:33.008578+0000 mgr.vm07.yrfcuj (mgr.14201) 605 : audit [DBG] from='client.15632 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:34.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:34 vm07 bash[17031]: audit 2026-03-07T07:05:33.008578+0000 mgr.vm07.yrfcuj (mgr.14201) 605 : audit [DBG] from='client.15632 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:34.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:34 vm07 bash[17031]: audit 2026-03-07T07:05:33.208253+0000 mon.vm07 (mon.0) 1027 : audit [DBG] from='client.? 192.168.123.107:0/3372375714' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:34.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:34 vm07 bash[17031]: audit 2026-03-07T07:05:33.208253+0000 mon.vm07 (mon.0) 1027 : audit [DBG] from='client.? 192.168.123.107:0/3372375714' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:34 vm09 bash[21220]: audit 2026-03-07T07:05:32.840594+0000 mgr.vm07.yrfcuj (mgr.14201) 604 : audit [DBG] from='client.15628 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:34 vm09 bash[21220]: audit 2026-03-07T07:05:32.840594+0000 mgr.vm07.yrfcuj (mgr.14201) 604 : audit [DBG] from='client.15628 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:34 vm09 bash[21220]: audit 2026-03-07T07:05:33.008578+0000 mgr.vm07.yrfcuj (mgr.14201) 605 : audit [DBG] from='client.15632 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:34 vm09 bash[21220]: audit 2026-03-07T07:05:33.008578+0000 mgr.vm07.yrfcuj (mgr.14201) 605 : audit [DBG] from='client.15632 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:34 vm09 bash[21220]: audit 2026-03-07T07:05:33.208253+0000 mon.vm07 (mon.0) 1027 : audit [DBG] from='client.? 192.168.123.107:0/3372375714' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:34 vm09 bash[21220]: audit 2026-03-07T07:05:33.208253+0000 mon.vm07 (mon.0) 1027 : audit [DBG] from='client.? 192.168.123.107:0/3372375714' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:35.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:35 vm09 bash[21220]: cluster 2026-03-07T07:05:33.823161+0000 mgr.vm07.yrfcuj (mgr.14201) 606 : cluster [DBG] pgmap v361: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:35.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:35 vm09 bash[21220]: cluster 2026-03-07T07:05:33.823161+0000 mgr.vm07.yrfcuj (mgr.14201) 606 : cluster [DBG] pgmap v361: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:35.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:35 vm07 bash[17031]: cluster 2026-03-07T07:05:33.823161+0000 mgr.vm07.yrfcuj (mgr.14201) 606 : cluster [DBG] pgmap v361: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:35.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:35 vm07 bash[17031]: cluster 2026-03-07T07:05:33.823161+0000 mgr.vm07.yrfcuj (mgr.14201) 606 : cluster [DBG] pgmap v361: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:36.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:36 vm07 bash[17031]: audit 2026-03-07T07:05:35.855919+0000 mon.vm07 (mon.0) 1028 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:05:36.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:36 vm07 bash[17031]: audit 2026-03-07T07:05:35.855919+0000 mon.vm07 (mon.0) 1028 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:05:36.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:36 vm09 bash[21220]: audit 2026-03-07T07:05:35.855919+0000 mon.vm07 (mon.0) 1028 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:05:36.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:36 vm09 bash[21220]: audit 2026-03-07T07:05:35.855919+0000 mon.vm07 (mon.0) 1028 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:05:37.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:37 vm07 bash[17031]: cluster 2026-03-07T07:05:35.823585+0000 mgr.vm07.yrfcuj (mgr.14201) 607 : cluster [DBG] pgmap v362: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:37.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:37 vm07 bash[17031]: cluster 2026-03-07T07:05:35.823585+0000 mgr.vm07.yrfcuj (mgr.14201) 607 : cluster [DBG] pgmap v362: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:37.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:37 vm07 bash[17031]: audit 2026-03-07T07:05:36.212420+0000 mon.vm07 (mon.0) 1029 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:05:37.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:37 vm07 bash[17031]: audit 2026-03-07T07:05:36.212420+0000 mon.vm07 (mon.0) 1029 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:05:37.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:37 vm07 bash[17031]: audit 2026-03-07T07:05:36.213039+0000 mon.vm07 (mon.0) 1030 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:05:37.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:37 vm07 bash[17031]: audit 2026-03-07T07:05:36.213039+0000 mon.vm07 (mon.0) 1030 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:05:37.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:37 vm07 bash[17031]: cluster 2026-03-07T07:05:36.214178+0000 mgr.vm07.yrfcuj (mgr.14201) 608 : cluster [DBG] pgmap v363: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:37.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:37 vm07 bash[17031]: cluster 2026-03-07T07:05:36.214178+0000 mgr.vm07.yrfcuj (mgr.14201) 608 : cluster [DBG] pgmap v363: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:37.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:37 vm07 bash[17031]: audit 2026-03-07T07:05:36.217777+0000 mon.vm07 (mon.0) 1031 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:05:37.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:37 vm07 bash[17031]: audit 2026-03-07T07:05:36.217777+0000 mon.vm07 (mon.0) 1031 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:05:37.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:37 vm07 bash[17031]: audit 2026-03-07T07:05:36.219162+0000 mon.vm07 (mon.0) 1032 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:05:37.401 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:37 vm07 bash[17031]: audit 2026-03-07T07:05:36.219162+0000 mon.vm07 (mon.0) 1032 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:05:37.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:37 vm09 bash[21220]: cluster 2026-03-07T07:05:35.823585+0000 mgr.vm07.yrfcuj (mgr.14201) 607 : cluster [DBG] pgmap v362: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:37.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:37 vm09 bash[21220]: cluster 2026-03-07T07:05:35.823585+0000 mgr.vm07.yrfcuj (mgr.14201) 607 : cluster [DBG] pgmap v362: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:37.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:37 vm09 bash[21220]: audit 2026-03-07T07:05:36.212420+0000 mon.vm07 (mon.0) 1029 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:05:37.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:37 vm09 bash[21220]: audit 2026-03-07T07:05:36.212420+0000 mon.vm07 (mon.0) 1029 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:05:37.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:37 vm09 bash[21220]: audit 2026-03-07T07:05:36.213039+0000 mon.vm07 (mon.0) 1030 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:05:37.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:37 vm09 bash[21220]: audit 2026-03-07T07:05:36.213039+0000 mon.vm07 (mon.0) 1030 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:05:37.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:37 vm09 bash[21220]: cluster 2026-03-07T07:05:36.214178+0000 mgr.vm07.yrfcuj (mgr.14201) 608 : cluster [DBG] pgmap v363: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:37.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:37 vm09 bash[21220]: cluster 2026-03-07T07:05:36.214178+0000 mgr.vm07.yrfcuj (mgr.14201) 608 : cluster [DBG] pgmap v363: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:37.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:37 vm09 bash[21220]: audit 2026-03-07T07:05:36.217777+0000 mon.vm07 (mon.0) 1031 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:05:37.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:37 vm09 bash[21220]: audit 2026-03-07T07:05:36.217777+0000 mon.vm07 (mon.0) 1031 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:05:37.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:37 vm09 bash[21220]: audit 2026-03-07T07:05:36.219162+0000 mon.vm07 (mon.0) 1032 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:05:37.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:37 vm09 bash[21220]: audit 2026-03-07T07:05:36.219162+0000 mon.vm07 (mon.0) 1032 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:05:38.384 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:05:38.537 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:05:38.537 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (2m) 86s ago 8m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:05:38.537 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (91s) 86s ago 7m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:05:38.538 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 62s ago 8m - - 2026-03-07T08:05:38.538 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (7m) 62s ago 7m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:05:38.742 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:05:38.742 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:05:38.742 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:05:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:39 vm09 bash[21220]: cluster 2026-03-07T07:05:38.214699+0000 mgr.vm07.yrfcuj (mgr.14201) 609 : cluster [DBG] pgmap v364: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:39 vm09 bash[21220]: cluster 2026-03-07T07:05:38.214699+0000 mgr.vm07.yrfcuj (mgr.14201) 609 : cluster [DBG] pgmap v364: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:39 vm09 bash[21220]: audit 2026-03-07T07:05:38.370944+0000 mgr.vm07.yrfcuj (mgr.14201) 610 : audit [DBG] from='client.25057 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:39 vm09 bash[21220]: audit 2026-03-07T07:05:38.370944+0000 mgr.vm07.yrfcuj (mgr.14201) 610 : audit [DBG] from='client.25057 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:39 vm09 bash[21220]: audit 2026-03-07T07:05:38.536750+0000 mgr.vm07.yrfcuj (mgr.14201) 611 : audit [DBG] from='client.15644 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:39 vm09 bash[21220]: audit 2026-03-07T07:05:38.536750+0000 mgr.vm07.yrfcuj (mgr.14201) 611 : audit [DBG] from='client.15644 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:39 vm09 bash[21220]: audit 2026-03-07T07:05:38.743402+0000 mon.vm07 (mon.0) 1033 : audit [DBG] from='client.? 192.168.123.107:0/3580011259' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:39 vm09 bash[21220]: audit 2026-03-07T07:05:38.743402+0000 mon.vm07 (mon.0) 1033 : audit [DBG] from='client.? 192.168.123.107:0/3580011259' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:39.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:39 vm07 bash[17031]: cluster 2026-03-07T07:05:38.214699+0000 mgr.vm07.yrfcuj (mgr.14201) 609 : cluster [DBG] pgmap v364: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:39.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:39 vm07 bash[17031]: cluster 2026-03-07T07:05:38.214699+0000 mgr.vm07.yrfcuj (mgr.14201) 609 : cluster [DBG] pgmap v364: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:39.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:39 vm07 bash[17031]: audit 2026-03-07T07:05:38.370944+0000 mgr.vm07.yrfcuj (mgr.14201) 610 : audit [DBG] from='client.25057 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:39.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:39 vm07 bash[17031]: audit 2026-03-07T07:05:38.370944+0000 mgr.vm07.yrfcuj (mgr.14201) 610 : audit [DBG] from='client.25057 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:39.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:39 vm07 bash[17031]: audit 2026-03-07T07:05:38.536750+0000 mgr.vm07.yrfcuj (mgr.14201) 611 : audit [DBG] from='client.15644 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:39.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:39 vm07 bash[17031]: audit 2026-03-07T07:05:38.536750+0000 mgr.vm07.yrfcuj (mgr.14201) 611 : audit [DBG] from='client.15644 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:39.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:39 vm07 bash[17031]: audit 2026-03-07T07:05:38.743402+0000 mon.vm07 (mon.0) 1033 : audit [DBG] from='client.? 192.168.123.107:0/3580011259' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:39.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:39 vm07 bash[17031]: audit 2026-03-07T07:05:38.743402+0000 mon.vm07 (mon.0) 1033 : audit [DBG] from='client.? 192.168.123.107:0/3580011259' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:40.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:40 vm09 bash[21220]: audit 2026-03-07T07:05:40.136907+0000 mon.vm07 (mon.0) 1034 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:05:40.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:40 vm09 bash[21220]: audit 2026-03-07T07:05:40.136907+0000 mon.vm07 (mon.0) 1034 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:05:40.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:40 vm07 bash[17031]: audit 2026-03-07T07:05:40.136907+0000 mon.vm07 (mon.0) 1034 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:05:40.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:40 vm07 bash[17031]: audit 2026-03-07T07:05:40.136907+0000 mon.vm07 (mon.0) 1034 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:05:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:41 vm09 bash[21220]: cluster 2026-03-07T07:05:40.215075+0000 mgr.vm07.yrfcuj (mgr.14201) 612 : cluster [DBG] pgmap v365: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:41 vm09 bash[21220]: cluster 2026-03-07T07:05:40.215075+0000 mgr.vm07.yrfcuj (mgr.14201) 612 : cluster [DBG] pgmap v365: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:41.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:41 vm07 bash[17031]: cluster 2026-03-07T07:05:40.215075+0000 mgr.vm07.yrfcuj (mgr.14201) 612 : cluster [DBG] pgmap v365: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:41.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:41 vm07 bash[17031]: cluster 2026-03-07T07:05:40.215075+0000 mgr.vm07.yrfcuj (mgr.14201) 612 : cluster [DBG] pgmap v365: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:43.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:43 vm09 bash[21220]: cluster 2026-03-07T07:05:42.215452+0000 mgr.vm07.yrfcuj (mgr.14201) 613 : cluster [DBG] pgmap v366: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:43.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:43 vm09 bash[21220]: cluster 2026-03-07T07:05:42.215452+0000 mgr.vm07.yrfcuj (mgr.14201) 613 : cluster [DBG] pgmap v366: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:43.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:43 vm07 bash[17031]: cluster 2026-03-07T07:05:42.215452+0000 mgr.vm07.yrfcuj (mgr.14201) 613 : cluster [DBG] pgmap v366: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:43.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:43 vm07 bash[17031]: cluster 2026-03-07T07:05:42.215452+0000 mgr.vm07.yrfcuj (mgr.14201) 613 : cluster [DBG] pgmap v366: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:43.906 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:05:44.052 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:05:44.052 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (2m) 92s ago 8m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:05:44.052 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (96s) 92s ago 8m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:05:44.052 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 68s ago 8m - - 2026-03-07T08:05:44.052 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (8m) 68s ago 8m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:05:44.236 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:05:44.236 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:05:44.236 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:05:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:44 vm09 bash[21220]: audit 2026-03-07T07:05:44.234503+0000 mon.vm09 (mon.1) 41 : audit [DBG] from='client.? 192.168.123.107:0/976651081' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:44 vm09 bash[21220]: audit 2026-03-07T07:05:44.234503+0000 mon.vm09 (mon.1) 41 : audit [DBG] from='client.? 192.168.123.107:0/976651081' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:44.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:44 vm07 bash[17031]: audit 2026-03-07T07:05:44.234503+0000 mon.vm09 (mon.1) 41 : audit [DBG] from='client.? 192.168.123.107:0/976651081' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:44.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:44 vm07 bash[17031]: audit 2026-03-07T07:05:44.234503+0000 mon.vm09 (mon.1) 41 : audit [DBG] from='client.? 192.168.123.107:0/976651081' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:45 vm09 bash[21220]: audit 2026-03-07T07:05:43.894114+0000 mgr.vm07.yrfcuj (mgr.14201) 614 : audit [DBG] from='client.15652 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:45 vm09 bash[21220]: audit 2026-03-07T07:05:43.894114+0000 mgr.vm07.yrfcuj (mgr.14201) 614 : audit [DBG] from='client.15652 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:45 vm09 bash[21220]: audit 2026-03-07T07:05:44.051655+0000 mgr.vm07.yrfcuj (mgr.14201) 615 : audit [DBG] from='client.25073 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:45 vm09 bash[21220]: audit 2026-03-07T07:05:44.051655+0000 mgr.vm07.yrfcuj (mgr.14201) 615 : audit [DBG] from='client.25073 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:45 vm09 bash[21220]: cluster 2026-03-07T07:05:44.215884+0000 mgr.vm07.yrfcuj (mgr.14201) 616 : cluster [DBG] pgmap v367: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:45 vm09 bash[21220]: cluster 2026-03-07T07:05:44.215884+0000 mgr.vm07.yrfcuj (mgr.14201) 616 : cluster [DBG] pgmap v367: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:45.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:45 vm07 bash[17031]: audit 2026-03-07T07:05:43.894114+0000 mgr.vm07.yrfcuj (mgr.14201) 614 : audit [DBG] from='client.15652 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:45.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:45 vm07 bash[17031]: audit 2026-03-07T07:05:43.894114+0000 mgr.vm07.yrfcuj (mgr.14201) 614 : audit [DBG] from='client.15652 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:45.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:45 vm07 bash[17031]: audit 2026-03-07T07:05:44.051655+0000 mgr.vm07.yrfcuj (mgr.14201) 615 : audit [DBG] from='client.25073 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:45.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:45 vm07 bash[17031]: audit 2026-03-07T07:05:44.051655+0000 mgr.vm07.yrfcuj (mgr.14201) 615 : audit [DBG] from='client.25073 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:45.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:45 vm07 bash[17031]: cluster 2026-03-07T07:05:44.215884+0000 mgr.vm07.yrfcuj (mgr.14201) 616 : cluster [DBG] pgmap v367: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:45.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:45 vm07 bash[17031]: cluster 2026-03-07T07:05:44.215884+0000 mgr.vm07.yrfcuj (mgr.14201) 616 : cluster [DBG] pgmap v367: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:47.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:47 vm09 bash[21220]: cluster 2026-03-07T07:05:46.216252+0000 mgr.vm07.yrfcuj (mgr.14201) 617 : cluster [DBG] pgmap v368: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:47.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:47 vm09 bash[21220]: cluster 2026-03-07T07:05:46.216252+0000 mgr.vm07.yrfcuj (mgr.14201) 617 : cluster [DBG] pgmap v368: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:47.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:47 vm07 bash[17031]: cluster 2026-03-07T07:05:46.216252+0000 mgr.vm07.yrfcuj (mgr.14201) 617 : cluster [DBG] pgmap v368: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:47.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:47 vm07 bash[17031]: cluster 2026-03-07T07:05:46.216252+0000 mgr.vm07.yrfcuj (mgr.14201) 617 : cluster [DBG] pgmap v368: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:05:49.406 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:05:49.558 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:05:49.558 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (2m) 97s ago 8m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:05:49.558 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (102s) 97s ago 8m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:05:49.558 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 73s ago 8m - - 2026-03-07T08:05:49.558 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (8m) 73s ago 8m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:05:49.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:49 vm09 bash[21220]: cluster 2026-03-07T07:05:48.216681+0000 mgr.vm07.yrfcuj (mgr.14201) 618 : cluster [DBG] pgmap v369: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:49.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:49 vm09 bash[21220]: cluster 2026-03-07T07:05:48.216681+0000 mgr.vm07.yrfcuj (mgr.14201) 618 : cluster [DBG] pgmap v369: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:49.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:49 vm07 bash[17031]: cluster 2026-03-07T07:05:48.216681+0000 mgr.vm07.yrfcuj (mgr.14201) 618 : cluster [DBG] pgmap v369: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:49.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:49 vm07 bash[17031]: cluster 2026-03-07T07:05:48.216681+0000 mgr.vm07.yrfcuj (mgr.14201) 618 : cluster [DBG] pgmap v369: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:49.739 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:05:49.739 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:05:49.739 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:05:50.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:50 vm07 bash[17031]: audit 2026-03-07T07:05:49.393873+0000 mgr.vm07.yrfcuj (mgr.14201) 619 : audit [DBG] from='client.15664 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:50.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:50 vm07 bash[17031]: audit 2026-03-07T07:05:49.393873+0000 mgr.vm07.yrfcuj (mgr.14201) 619 : audit [DBG] from='client.15664 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:50.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:50 vm07 bash[17031]: audit 2026-03-07T07:05:49.557347+0000 mgr.vm07.yrfcuj (mgr.14201) 620 : audit [DBG] from='client.15668 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:50.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:50 vm07 bash[17031]: audit 2026-03-07T07:05:49.557347+0000 mgr.vm07.yrfcuj (mgr.14201) 620 : audit [DBG] from='client.15668 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:50.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:50 vm07 bash[17031]: audit 2026-03-07T07:05:49.740949+0000 mon.vm07 (mon.0) 1035 : audit [DBG] from='client.? 192.168.123.107:0/668574823' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:50.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:50 vm07 bash[17031]: audit 2026-03-07T07:05:49.740949+0000 mon.vm07 (mon.0) 1035 : audit [DBG] from='client.? 192.168.123.107:0/668574823' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:50 vm09 bash[21220]: audit 2026-03-07T07:05:49.393873+0000 mgr.vm07.yrfcuj (mgr.14201) 619 : audit [DBG] from='client.15664 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:50 vm09 bash[21220]: audit 2026-03-07T07:05:49.393873+0000 mgr.vm07.yrfcuj (mgr.14201) 619 : audit [DBG] from='client.15664 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:50 vm09 bash[21220]: audit 2026-03-07T07:05:49.557347+0000 mgr.vm07.yrfcuj (mgr.14201) 620 : audit [DBG] from='client.15668 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:50 vm09 bash[21220]: audit 2026-03-07T07:05:49.557347+0000 mgr.vm07.yrfcuj (mgr.14201) 620 : audit [DBG] from='client.15668 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:50 vm09 bash[21220]: audit 2026-03-07T07:05:49.740949+0000 mon.vm07 (mon.0) 1035 : audit [DBG] from='client.? 192.168.123.107:0/668574823' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:50 vm09 bash[21220]: audit 2026-03-07T07:05:49.740949+0000 mon.vm07 (mon.0) 1035 : audit [DBG] from='client.? 192.168.123.107:0/668574823' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:51.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:51 vm07 bash[17031]: cluster 2026-03-07T07:05:50.217020+0000 mgr.vm07.yrfcuj (mgr.14201) 621 : cluster [DBG] pgmap v370: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:51.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:51 vm07 bash[17031]: cluster 2026-03-07T07:05:50.217020+0000 mgr.vm07.yrfcuj (mgr.14201) 621 : cluster [DBG] pgmap v370: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:51 vm09 bash[21220]: cluster 2026-03-07T07:05:50.217020+0000 mgr.vm07.yrfcuj (mgr.14201) 621 : cluster [DBG] pgmap v370: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:51 vm09 bash[21220]: cluster 2026-03-07T07:05:50.217020+0000 mgr.vm07.yrfcuj (mgr.14201) 621 : cluster [DBG] pgmap v370: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:53.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:53 vm07 bash[17031]: cluster 2026-03-07T07:05:52.217397+0000 mgr.vm07.yrfcuj (mgr.14201) 622 : cluster [DBG] pgmap v371: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:53.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:53 vm07 bash[17031]: cluster 2026-03-07T07:05:52.217397+0000 mgr.vm07.yrfcuj (mgr.14201) 622 : cluster [DBG] pgmap v371: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:53.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:53 vm09 bash[21220]: cluster 2026-03-07T07:05:52.217397+0000 mgr.vm07.yrfcuj (mgr.14201) 622 : cluster [DBG] pgmap v371: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:53.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:53 vm09 bash[21220]: cluster 2026-03-07T07:05:52.217397+0000 mgr.vm07.yrfcuj (mgr.14201) 622 : cluster [DBG] pgmap v371: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:05:54.907 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:05:55.054 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:05:55.054 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (2m) 103s ago 8m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:05:55.054 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (107s) 103s ago 8m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:05:55.054 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 79s ago 8m - - 2026-03-07T08:05:55.054 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (8m) 79s ago 8m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:05:55.240 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:05:55.241 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:05:55.241 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:05:55.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:55 vm07 bash[17031]: cluster 2026-03-07T07:05:54.217745+0000 mgr.vm07.yrfcuj (mgr.14201) 623 : cluster [DBG] pgmap v372: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:55.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:55 vm07 bash[17031]: cluster 2026-03-07T07:05:54.217745+0000 mgr.vm07.yrfcuj (mgr.14201) 623 : cluster [DBG] pgmap v372: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:55.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:55 vm07 bash[17031]: audit 2026-03-07T07:05:55.137073+0000 mon.vm07 (mon.0) 1036 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:05:55.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:55 vm07 bash[17031]: audit 2026-03-07T07:05:55.137073+0000 mon.vm07 (mon.0) 1036 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:05:55.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:55 vm07 bash[17031]: audit 2026-03-07T07:05:55.242579+0000 mon.vm07 (mon.0) 1037 : audit [DBG] from='client.? 192.168.123.107:0/2823488522' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:55.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:55 vm07 bash[17031]: audit 2026-03-07T07:05:55.242579+0000 mon.vm07 (mon.0) 1037 : audit [DBG] from='client.? 192.168.123.107:0/2823488522' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:55 vm09 bash[21220]: cluster 2026-03-07T07:05:54.217745+0000 mgr.vm07.yrfcuj (mgr.14201) 623 : cluster [DBG] pgmap v372: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:55 vm09 bash[21220]: cluster 2026-03-07T07:05:54.217745+0000 mgr.vm07.yrfcuj (mgr.14201) 623 : cluster [DBG] pgmap v372: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:55 vm09 bash[21220]: audit 2026-03-07T07:05:55.137073+0000 mon.vm07 (mon.0) 1036 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:05:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:55 vm09 bash[21220]: audit 2026-03-07T07:05:55.137073+0000 mon.vm07 (mon.0) 1036 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:05:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:55 vm09 bash[21220]: audit 2026-03-07T07:05:55.242579+0000 mon.vm07 (mon.0) 1037 : audit [DBG] from='client.? 192.168.123.107:0/2823488522' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:55 vm09 bash[21220]: audit 2026-03-07T07:05:55.242579+0000 mon.vm07 (mon.0) 1037 : audit [DBG] from='client.? 192.168.123.107:0/2823488522' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:05:56.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:56 vm07 bash[17031]: audit 2026-03-07T07:05:54.894770+0000 mgr.vm07.yrfcuj (mgr.14201) 624 : audit [DBG] from='client.15676 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:56.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:56 vm07 bash[17031]: audit 2026-03-07T07:05:54.894770+0000 mgr.vm07.yrfcuj (mgr.14201) 624 : audit [DBG] from='client.15676 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:56.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:56 vm07 bash[17031]: audit 2026-03-07T07:05:55.053838+0000 mgr.vm07.yrfcuj (mgr.14201) 625 : audit [DBG] from='client.15680 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:56.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:56 vm07 bash[17031]: audit 2026-03-07T07:05:55.053838+0000 mgr.vm07.yrfcuj (mgr.14201) 625 : audit [DBG] from='client.15680 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:56.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:56 vm09 bash[21220]: audit 2026-03-07T07:05:54.894770+0000 mgr.vm07.yrfcuj (mgr.14201) 624 : audit [DBG] from='client.15676 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:56.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:56 vm09 bash[21220]: audit 2026-03-07T07:05:54.894770+0000 mgr.vm07.yrfcuj (mgr.14201) 624 : audit [DBG] from='client.15676 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:56.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:56 vm09 bash[21220]: audit 2026-03-07T07:05:55.053838+0000 mgr.vm07.yrfcuj (mgr.14201) 625 : audit [DBG] from='client.15680 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:56.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:56 vm09 bash[21220]: audit 2026-03-07T07:05:55.053838+0000 mgr.vm07.yrfcuj (mgr.14201) 625 : audit [DBG] from='client.15680 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:05:57.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:57 vm07 bash[17031]: cluster 2026-03-07T07:05:56.218108+0000 mgr.vm07.yrfcuj (mgr.14201) 626 : cluster [DBG] pgmap v373: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:57.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:57 vm07 bash[17031]: cluster 2026-03-07T07:05:56.218108+0000 mgr.vm07.yrfcuj (mgr.14201) 626 : cluster [DBG] pgmap v373: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:57.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:57 vm09 bash[21220]: cluster 2026-03-07T07:05:56.218108+0000 mgr.vm07.yrfcuj (mgr.14201) 626 : cluster [DBG] pgmap v373: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:57.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:57 vm09 bash[21220]: cluster 2026-03-07T07:05:56.218108+0000 mgr.vm07.yrfcuj (mgr.14201) 626 : cluster [DBG] pgmap v373: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:59.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:59 vm09 bash[21220]: cluster 2026-03-07T07:05:58.218524+0000 mgr.vm07.yrfcuj (mgr.14201) 627 : cluster [DBG] pgmap v374: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:59.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:05:59 vm09 bash[21220]: cluster 2026-03-07T07:05:58.218524+0000 mgr.vm07.yrfcuj (mgr.14201) 627 : cluster [DBG] pgmap v374: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:59.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:59 vm07 bash[17031]: cluster 2026-03-07T07:05:58.218524+0000 mgr.vm07.yrfcuj (mgr.14201) 627 : cluster [DBG] pgmap v374: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:05:59.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:05:59 vm07 bash[17031]: cluster 2026-03-07T07:05:58.218524+0000 mgr.vm07.yrfcuj (mgr.14201) 627 : cluster [DBG] pgmap v374: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:00.411 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:06:00.619 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:06:00.619 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (2m) 108s ago 8m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:06:00.619 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (113s) 108s ago 8m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:06:00.619 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 84s ago 8m - - 2026-03-07T08:06:00.619 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (8m) 84s ago 8m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:06:00.805 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:06:00.805 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:06:00.805 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:06:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:01 vm09 bash[21220]: cluster 2026-03-07T07:06:00.218889+0000 mgr.vm07.yrfcuj (mgr.14201) 628 : cluster [DBG] pgmap v375: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:06:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:01 vm09 bash[21220]: cluster 2026-03-07T07:06:00.218889+0000 mgr.vm07.yrfcuj (mgr.14201) 628 : cluster [DBG] pgmap v375: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:06:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:01 vm09 bash[21220]: audit 2026-03-07T07:06:00.397094+0000 mgr.vm07.yrfcuj (mgr.14201) 629 : audit [DBG] from='client.15688 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:01 vm09 bash[21220]: audit 2026-03-07T07:06:00.397094+0000 mgr.vm07.yrfcuj (mgr.14201) 629 : audit [DBG] from='client.15688 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:01 vm09 bash[21220]: audit 2026-03-07T07:06:00.806918+0000 mon.vm07 (mon.0) 1038 : audit [DBG] from='client.? 192.168.123.107:0/2692359129' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:01 vm09 bash[21220]: audit 2026-03-07T07:06:00.806918+0000 mon.vm07 (mon.0) 1038 : audit [DBG] from='client.? 192.168.123.107:0/2692359129' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:01.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:01 vm07 bash[17031]: cluster 2026-03-07T07:06:00.218889+0000 mgr.vm07.yrfcuj (mgr.14201) 628 : cluster [DBG] pgmap v375: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:06:01.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:01 vm07 bash[17031]: cluster 2026-03-07T07:06:00.218889+0000 mgr.vm07.yrfcuj (mgr.14201) 628 : cluster [DBG] pgmap v375: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:06:01.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:01 vm07 bash[17031]: audit 2026-03-07T07:06:00.397094+0000 mgr.vm07.yrfcuj (mgr.14201) 629 : audit [DBG] from='client.15688 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:01.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:01 vm07 bash[17031]: audit 2026-03-07T07:06:00.397094+0000 mgr.vm07.yrfcuj (mgr.14201) 629 : audit [DBG] from='client.15688 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:01.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:01 vm07 bash[17031]: audit 2026-03-07T07:06:00.806918+0000 mon.vm07 (mon.0) 1038 : audit [DBG] from='client.? 192.168.123.107:0/2692359129' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:01.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:01 vm07 bash[17031]: audit 2026-03-07T07:06:00.806918+0000 mon.vm07 (mon.0) 1038 : audit [DBG] from='client.? 192.168.123.107:0/2692359129' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:02.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:02 vm09 bash[21220]: audit 2026-03-07T07:06:00.618587+0000 mgr.vm07.yrfcuj (mgr.14201) 630 : audit [DBG] from='client.15692 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:02.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:02 vm09 bash[21220]: audit 2026-03-07T07:06:00.618587+0000 mgr.vm07.yrfcuj (mgr.14201) 630 : audit [DBG] from='client.15692 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:02.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:02 vm07 bash[17031]: audit 2026-03-07T07:06:00.618587+0000 mgr.vm07.yrfcuj (mgr.14201) 630 : audit [DBG] from='client.15692 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:02.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:02 vm07 bash[17031]: audit 2026-03-07T07:06:00.618587+0000 mgr.vm07.yrfcuj (mgr.14201) 630 : audit [DBG] from='client.15692 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:03.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:03 vm09 bash[21220]: cluster 2026-03-07T07:06:02.219239+0000 mgr.vm07.yrfcuj (mgr.14201) 631 : cluster [DBG] pgmap v376: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:03.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:03 vm09 bash[21220]: cluster 2026-03-07T07:06:02.219239+0000 mgr.vm07.yrfcuj (mgr.14201) 631 : cluster [DBG] pgmap v376: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:03.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:03 vm07 bash[17031]: cluster 2026-03-07T07:06:02.219239+0000 mgr.vm07.yrfcuj (mgr.14201) 631 : cluster [DBG] pgmap v376: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:03.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:03 vm07 bash[17031]: cluster 2026-03-07T07:06:02.219239+0000 mgr.vm07.yrfcuj (mgr.14201) 631 : cluster [DBG] pgmap v376: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:05 vm09 bash[21220]: cluster 2026-03-07T07:06:04.219665+0000 mgr.vm07.yrfcuj (mgr.14201) 632 : cluster [DBG] pgmap v377: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:05 vm09 bash[21220]: cluster 2026-03-07T07:06:04.219665+0000 mgr.vm07.yrfcuj (mgr.14201) 632 : cluster [DBG] pgmap v377: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:05.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:05 vm07 bash[17031]: cluster 2026-03-07T07:06:04.219665+0000 mgr.vm07.yrfcuj (mgr.14201) 632 : cluster [DBG] pgmap v377: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:05.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:05 vm07 bash[17031]: cluster 2026-03-07T07:06:04.219665+0000 mgr.vm07.yrfcuj (mgr.14201) 632 : cluster [DBG] pgmap v377: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:05.974 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:06:06.122 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:06:06.122 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (2m) 114s ago 8m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:06:06.122 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (118s) 114s ago 8m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:06:06.123 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 90s ago 8m - - 2026-03-07T08:06:06.123 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (8m) 90s ago 8m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:06:06.312 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:06:06.312 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:06:06.312 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:06:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:06 vm09 bash[21220]: audit 2026-03-07T07:06:06.313967+0000 mon.vm07 (mon.0) 1039 : audit [DBG] from='client.? 192.168.123.107:0/1905451530' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:06 vm09 bash[21220]: audit 2026-03-07T07:06:06.313967+0000 mon.vm07 (mon.0) 1039 : audit [DBG] from='client.? 192.168.123.107:0/1905451530' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:06.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:06 vm07 bash[17031]: audit 2026-03-07T07:06:06.313967+0000 mon.vm07 (mon.0) 1039 : audit [DBG] from='client.? 192.168.123.107:0/1905451530' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:06.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:06 vm07 bash[17031]: audit 2026-03-07T07:06:06.313967+0000 mon.vm07 (mon.0) 1039 : audit [DBG] from='client.? 192.168.123.107:0/1905451530' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:07.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:07 vm09 bash[21220]: audit 2026-03-07T07:06:05.964113+0000 mgr.vm07.yrfcuj (mgr.14201) 633 : audit [DBG] from='client.15700 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:07.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:07 vm09 bash[21220]: audit 2026-03-07T07:06:05.964113+0000 mgr.vm07.yrfcuj (mgr.14201) 633 : audit [DBG] from='client.15700 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:07.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:07 vm09 bash[21220]: audit 2026-03-07T07:06:06.122313+0000 mgr.vm07.yrfcuj (mgr.14201) 634 : audit [DBG] from='client.15704 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:07.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:07 vm09 bash[21220]: audit 2026-03-07T07:06:06.122313+0000 mgr.vm07.yrfcuj (mgr.14201) 634 : audit [DBG] from='client.15704 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:07.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:07 vm09 bash[21220]: cluster 2026-03-07T07:06:06.219941+0000 mgr.vm07.yrfcuj (mgr.14201) 635 : cluster [DBG] pgmap v378: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:07.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:07 vm09 bash[21220]: cluster 2026-03-07T07:06:06.219941+0000 mgr.vm07.yrfcuj (mgr.14201) 635 : cluster [DBG] pgmap v378: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:07.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:07 vm07 bash[17031]: audit 2026-03-07T07:06:05.964113+0000 mgr.vm07.yrfcuj (mgr.14201) 633 : audit [DBG] from='client.15700 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:07.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:07 vm07 bash[17031]: audit 2026-03-07T07:06:05.964113+0000 mgr.vm07.yrfcuj (mgr.14201) 633 : audit [DBG] from='client.15700 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:07.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:07 vm07 bash[17031]: audit 2026-03-07T07:06:06.122313+0000 mgr.vm07.yrfcuj (mgr.14201) 634 : audit [DBG] from='client.15704 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:07.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:07 vm07 bash[17031]: audit 2026-03-07T07:06:06.122313+0000 mgr.vm07.yrfcuj (mgr.14201) 634 : audit [DBG] from='client.15704 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:07.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:07 vm07 bash[17031]: cluster 2026-03-07T07:06:06.219941+0000 mgr.vm07.yrfcuj (mgr.14201) 635 : cluster [DBG] pgmap v378: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:07.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:07 vm07 bash[17031]: cluster 2026-03-07T07:06:06.219941+0000 mgr.vm07.yrfcuj (mgr.14201) 635 : cluster [DBG] pgmap v378: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:09.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:09 vm09 bash[21220]: cluster 2026-03-07T07:06:08.220395+0000 mgr.vm07.yrfcuj (mgr.14201) 636 : cluster [DBG] pgmap v379: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:09.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:09 vm09 bash[21220]: cluster 2026-03-07T07:06:08.220395+0000 mgr.vm07.yrfcuj (mgr.14201) 636 : cluster [DBG] pgmap v379: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:09.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:09 vm07 bash[17031]: cluster 2026-03-07T07:06:08.220395+0000 mgr.vm07.yrfcuj (mgr.14201) 636 : cluster [DBG] pgmap v379: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:09.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:09 vm07 bash[17031]: cluster 2026-03-07T07:06:08.220395+0000 mgr.vm07.yrfcuj (mgr.14201) 636 : cluster [DBG] pgmap v379: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:10 vm09 bash[21220]: audit 2026-03-07T07:06:10.137274+0000 mon.vm07 (mon.0) 1040 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:06:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:10 vm09 bash[21220]: audit 2026-03-07T07:06:10.137274+0000 mon.vm07 (mon.0) 1040 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:06:10.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:10 vm07 bash[17031]: audit 2026-03-07T07:06:10.137274+0000 mon.vm07 (mon.0) 1040 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:06:10.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:10 vm07 bash[17031]: audit 2026-03-07T07:06:10.137274+0000 mon.vm07 (mon.0) 1040 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:06:11.475 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:06:11.622 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:06:11.622 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (2m) 119s ago 8m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:06:11.622 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 119s ago 8m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:06:11.622 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 95s ago 8m - - 2026-03-07T08:06:11.622 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (8m) 95s ago 8m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:06:11.802 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:06:11.802 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:06:11.802 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:06:11.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:11 vm09 bash[21220]: cluster 2026-03-07T07:06:10.220768+0000 mgr.vm07.yrfcuj (mgr.14201) 637 : cluster [DBG] pgmap v380: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:11.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:11 vm09 bash[21220]: cluster 2026-03-07T07:06:10.220768+0000 mgr.vm07.yrfcuj (mgr.14201) 637 : cluster [DBG] pgmap v380: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:11.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:11 vm09 bash[21220]: audit 2026-03-07T07:06:11.462970+0000 mgr.vm07.yrfcuj (mgr.14201) 638 : audit [DBG] from='client.15712 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:11.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:11 vm09 bash[21220]: audit 2026-03-07T07:06:11.462970+0000 mgr.vm07.yrfcuj (mgr.14201) 638 : audit [DBG] from='client.15712 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:11.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:11 vm07 bash[17031]: cluster 2026-03-07T07:06:10.220768+0000 mgr.vm07.yrfcuj (mgr.14201) 637 : cluster [DBG] pgmap v380: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:11.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:11 vm07 bash[17031]: cluster 2026-03-07T07:06:10.220768+0000 mgr.vm07.yrfcuj (mgr.14201) 637 : cluster [DBG] pgmap v380: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:11.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:11 vm07 bash[17031]: audit 2026-03-07T07:06:11.462970+0000 mgr.vm07.yrfcuj (mgr.14201) 638 : audit [DBG] from='client.15712 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:11.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:11 vm07 bash[17031]: audit 2026-03-07T07:06:11.462970+0000 mgr.vm07.yrfcuj (mgr.14201) 638 : audit [DBG] from='client.15712 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:12.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:12 vm09 bash[21220]: audit 2026-03-07T07:06:11.621894+0000 mgr.vm07.yrfcuj (mgr.14201) 639 : audit [DBG] from='client.15716 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:12.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:12 vm09 bash[21220]: audit 2026-03-07T07:06:11.621894+0000 mgr.vm07.yrfcuj (mgr.14201) 639 : audit [DBG] from='client.15716 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:12.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:12 vm09 bash[21220]: audit 2026-03-07T07:06:11.804568+0000 mon.vm07 (mon.0) 1041 : audit [DBG] from='client.? 192.168.123.107:0/1586199275' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:12.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:12 vm09 bash[21220]: audit 2026-03-07T07:06:11.804568+0000 mon.vm07 (mon.0) 1041 : audit [DBG] from='client.? 192.168.123.107:0/1586199275' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:12.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:12 vm09 bash[21220]: cluster 2026-03-07T07:06:12.221235+0000 mgr.vm07.yrfcuj (mgr.14201) 640 : cluster [DBG] pgmap v381: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:12.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:12 vm09 bash[21220]: cluster 2026-03-07T07:06:12.221235+0000 mgr.vm07.yrfcuj (mgr.14201) 640 : cluster [DBG] pgmap v381: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:12.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:12 vm07 bash[17031]: audit 2026-03-07T07:06:11.621894+0000 mgr.vm07.yrfcuj (mgr.14201) 639 : audit [DBG] from='client.15716 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:12.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:12 vm07 bash[17031]: audit 2026-03-07T07:06:11.621894+0000 mgr.vm07.yrfcuj (mgr.14201) 639 : audit [DBG] from='client.15716 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:12.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:12 vm07 bash[17031]: audit 2026-03-07T07:06:11.804568+0000 mon.vm07 (mon.0) 1041 : audit [DBG] from='client.? 192.168.123.107:0/1586199275' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:12.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:12 vm07 bash[17031]: audit 2026-03-07T07:06:11.804568+0000 mon.vm07 (mon.0) 1041 : audit [DBG] from='client.? 192.168.123.107:0/1586199275' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:12.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:12 vm07 bash[17031]: cluster 2026-03-07T07:06:12.221235+0000 mgr.vm07.yrfcuj (mgr.14201) 640 : cluster [DBG] pgmap v381: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:12.901 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:12 vm07 bash[17031]: cluster 2026-03-07T07:06:12.221235+0000 mgr.vm07.yrfcuj (mgr.14201) 640 : cluster [DBG] pgmap v381: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:15 vm09 bash[21220]: cluster 2026-03-07T07:06:14.221676+0000 mgr.vm07.yrfcuj (mgr.14201) 641 : cluster [DBG] pgmap v382: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:15.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:15 vm09 bash[21220]: cluster 2026-03-07T07:06:14.221676+0000 mgr.vm07.yrfcuj (mgr.14201) 641 : cluster [DBG] pgmap v382: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:15.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:15 vm07 bash[17031]: cluster 2026-03-07T07:06:14.221676+0000 mgr.vm07.yrfcuj (mgr.14201) 641 : cluster [DBG] pgmap v382: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:15.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:15 vm07 bash[17031]: cluster 2026-03-07T07:06:14.221676+0000 mgr.vm07.yrfcuj (mgr.14201) 641 : cluster [DBG] pgmap v382: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:16.977 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:06:17.141 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:06:17.141 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (2m) 2m ago 8m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:06:17.141 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 2m ago 8m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:06:17.141 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 101s ago 8m - - 2026-03-07T08:06:17.141 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (8m) 101s ago 8m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:06:17.335 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:06:17.335 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:06:17.335 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:06:17.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:17 vm09 bash[21220]: cluster 2026-03-07T07:06:16.222049+0000 mgr.vm07.yrfcuj (mgr.14201) 642 : cluster [DBG] pgmap v383: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:17.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:17 vm09 bash[21220]: cluster 2026-03-07T07:06:16.222049+0000 mgr.vm07.yrfcuj (mgr.14201) 642 : cluster [DBG] pgmap v383: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:17.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:17 vm07 bash[17031]: cluster 2026-03-07T07:06:16.222049+0000 mgr.vm07.yrfcuj (mgr.14201) 642 : cluster [DBG] pgmap v383: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:17.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:17 vm07 bash[17031]: cluster 2026-03-07T07:06:16.222049+0000 mgr.vm07.yrfcuj (mgr.14201) 642 : cluster [DBG] pgmap v383: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:18 vm09 bash[21220]: audit 2026-03-07T07:06:16.965061+0000 mgr.vm07.yrfcuj (mgr.14201) 643 : audit [DBG] from='client.15724 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:18 vm09 bash[21220]: audit 2026-03-07T07:06:16.965061+0000 mgr.vm07.yrfcuj (mgr.14201) 643 : audit [DBG] from='client.15724 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:18 vm09 bash[21220]: audit 2026-03-07T07:06:17.140551+0000 mgr.vm07.yrfcuj (mgr.14201) 644 : audit [DBG] from='client.15728 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:18 vm09 bash[21220]: audit 2026-03-07T07:06:17.140551+0000 mgr.vm07.yrfcuj (mgr.14201) 644 : audit [DBG] from='client.15728 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:18 vm09 bash[21220]: audit 2026-03-07T07:06:17.337174+0000 mon.vm07 (mon.0) 1042 : audit [DBG] from='client.? 192.168.123.107:0/913984550' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:18 vm09 bash[21220]: audit 2026-03-07T07:06:17.337174+0000 mon.vm07 (mon.0) 1042 : audit [DBG] from='client.? 192.168.123.107:0/913984550' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:18.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:18 vm07 bash[17031]: audit 2026-03-07T07:06:16.965061+0000 mgr.vm07.yrfcuj (mgr.14201) 643 : audit [DBG] from='client.15724 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:18.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:18 vm07 bash[17031]: audit 2026-03-07T07:06:16.965061+0000 mgr.vm07.yrfcuj (mgr.14201) 643 : audit [DBG] from='client.15724 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:18.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:18 vm07 bash[17031]: audit 2026-03-07T07:06:17.140551+0000 mgr.vm07.yrfcuj (mgr.14201) 644 : audit [DBG] from='client.15728 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:18.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:18 vm07 bash[17031]: audit 2026-03-07T07:06:17.140551+0000 mgr.vm07.yrfcuj (mgr.14201) 644 : audit [DBG] from='client.15728 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:18.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:18 vm07 bash[17031]: audit 2026-03-07T07:06:17.337174+0000 mon.vm07 (mon.0) 1042 : audit [DBG] from='client.? 192.168.123.107:0/913984550' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:18.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:18 vm07 bash[17031]: audit 2026-03-07T07:06:17.337174+0000 mon.vm07 (mon.0) 1042 : audit [DBG] from='client.? 192.168.123.107:0/913984550' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:19.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:19 vm09 bash[21220]: cluster 2026-03-07T07:06:18.222497+0000 mgr.vm07.yrfcuj (mgr.14201) 645 : cluster [DBG] pgmap v384: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:19.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:19 vm09 bash[21220]: cluster 2026-03-07T07:06:18.222497+0000 mgr.vm07.yrfcuj (mgr.14201) 645 : cluster [DBG] pgmap v384: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:19.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:19 vm07 bash[17031]: cluster 2026-03-07T07:06:18.222497+0000 mgr.vm07.yrfcuj (mgr.14201) 645 : cluster [DBG] pgmap v384: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:19.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:19 vm07 bash[17031]: cluster 2026-03-07T07:06:18.222497+0000 mgr.vm07.yrfcuj (mgr.14201) 645 : cluster [DBG] pgmap v384: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:21.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:21 vm09 bash[21220]: cluster 2026-03-07T07:06:20.222854+0000 mgr.vm07.yrfcuj (mgr.14201) 646 : cluster [DBG] pgmap v385: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:21.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:21 vm09 bash[21220]: cluster 2026-03-07T07:06:20.222854+0000 mgr.vm07.yrfcuj (mgr.14201) 646 : cluster [DBG] pgmap v385: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:21.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:21 vm07 bash[17031]: cluster 2026-03-07T07:06:20.222854+0000 mgr.vm07.yrfcuj (mgr.14201) 646 : cluster [DBG] pgmap v385: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:21.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:21 vm07 bash[17031]: cluster 2026-03-07T07:06:20.222854+0000 mgr.vm07.yrfcuj (mgr.14201) 646 : cluster [DBG] pgmap v385: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:22.499 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:06:22.644 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:06:22.644 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (2m) 2m ago 8m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:06:22.644 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 2m ago 8m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:06:22.644 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 106s ago 8m - - 2026-03-07T08:06:22.644 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (8m) 106s ago 8m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:06:22.822 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:06:22.822 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:06:22.822 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:06:23.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:23 vm09 bash[21220]: cluster 2026-03-07T07:06:22.223250+0000 mgr.vm07.yrfcuj (mgr.14201) 647 : cluster [DBG] pgmap v386: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:23.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:23 vm09 bash[21220]: cluster 2026-03-07T07:06:22.223250+0000 mgr.vm07.yrfcuj (mgr.14201) 647 : cluster [DBG] pgmap v386: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:23.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:23 vm09 bash[21220]: audit 2026-03-07T07:06:22.487754+0000 mgr.vm07.yrfcuj (mgr.14201) 648 : audit [DBG] from='client.15736 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:23.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:23 vm09 bash[21220]: audit 2026-03-07T07:06:22.487754+0000 mgr.vm07.yrfcuj (mgr.14201) 648 : audit [DBG] from='client.15736 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:23.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:23 vm09 bash[21220]: audit 2026-03-07T07:06:22.824463+0000 mon.vm07 (mon.0) 1043 : audit [DBG] from='client.? 192.168.123.107:0/3809916651' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:23.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:23 vm09 bash[21220]: audit 2026-03-07T07:06:22.824463+0000 mon.vm07 (mon.0) 1043 : audit [DBG] from='client.? 192.168.123.107:0/3809916651' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:23.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:23 vm07 bash[17031]: cluster 2026-03-07T07:06:22.223250+0000 mgr.vm07.yrfcuj (mgr.14201) 647 : cluster [DBG] pgmap v386: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:23.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:23 vm07 bash[17031]: cluster 2026-03-07T07:06:22.223250+0000 mgr.vm07.yrfcuj (mgr.14201) 647 : cluster [DBG] pgmap v386: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:23.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:23 vm07 bash[17031]: audit 2026-03-07T07:06:22.487754+0000 mgr.vm07.yrfcuj (mgr.14201) 648 : audit [DBG] from='client.15736 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:23.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:23 vm07 bash[17031]: audit 2026-03-07T07:06:22.487754+0000 mgr.vm07.yrfcuj (mgr.14201) 648 : audit [DBG] from='client.15736 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:23.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:23 vm07 bash[17031]: audit 2026-03-07T07:06:22.824463+0000 mon.vm07 (mon.0) 1043 : audit [DBG] from='client.? 192.168.123.107:0/3809916651' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:23.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:23 vm07 bash[17031]: audit 2026-03-07T07:06:22.824463+0000 mon.vm07 (mon.0) 1043 : audit [DBG] from='client.? 192.168.123.107:0/3809916651' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:24 vm09 bash[21220]: audit 2026-03-07T07:06:22.643289+0000 mgr.vm07.yrfcuj (mgr.14201) 649 : audit [DBG] from='client.15740 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:24 vm09 bash[21220]: audit 2026-03-07T07:06:22.643289+0000 mgr.vm07.yrfcuj (mgr.14201) 649 : audit [DBG] from='client.15740 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:24.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:24 vm07 bash[17031]: audit 2026-03-07T07:06:22.643289+0000 mgr.vm07.yrfcuj (mgr.14201) 649 : audit [DBG] from='client.15740 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:24.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:24 vm07 bash[17031]: audit 2026-03-07T07:06:22.643289+0000 mgr.vm07.yrfcuj (mgr.14201) 649 : audit [DBG] from='client.15740 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:25 vm09 bash[21220]: cluster 2026-03-07T07:06:24.223813+0000 mgr.vm07.yrfcuj (mgr.14201) 650 : cluster [DBG] pgmap v387: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:25 vm09 bash[21220]: cluster 2026-03-07T07:06:24.223813+0000 mgr.vm07.yrfcuj (mgr.14201) 650 : cluster [DBG] pgmap v387: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:25 vm09 bash[21220]: audit 2026-03-07T07:06:25.137548+0000 mon.vm07 (mon.0) 1044 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:06:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:25 vm09 bash[21220]: audit 2026-03-07T07:06:25.137548+0000 mon.vm07 (mon.0) 1044 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:06:25.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:25 vm07 bash[17031]: cluster 2026-03-07T07:06:24.223813+0000 mgr.vm07.yrfcuj (mgr.14201) 650 : cluster [DBG] pgmap v387: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:25.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:25 vm07 bash[17031]: cluster 2026-03-07T07:06:24.223813+0000 mgr.vm07.yrfcuj (mgr.14201) 650 : cluster [DBG] pgmap v387: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:25.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:25 vm07 bash[17031]: audit 2026-03-07T07:06:25.137548+0000 mon.vm07 (mon.0) 1044 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:06:25.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:25 vm07 bash[17031]: audit 2026-03-07T07:06:25.137548+0000 mon.vm07 (mon.0) 1044 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:06:27.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:27 vm09 bash[21220]: cluster 2026-03-07T07:06:26.224309+0000 mgr.vm07.yrfcuj (mgr.14201) 651 : cluster [DBG] pgmap v388: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:27.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:27 vm09 bash[21220]: cluster 2026-03-07T07:06:26.224309+0000 mgr.vm07.yrfcuj (mgr.14201) 651 : cluster [DBG] pgmap v388: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:27.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:27 vm07 bash[17031]: cluster 2026-03-07T07:06:26.224309+0000 mgr.vm07.yrfcuj (mgr.14201) 651 : cluster [DBG] pgmap v388: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:27.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:27 vm07 bash[17031]: cluster 2026-03-07T07:06:26.224309+0000 mgr.vm07.yrfcuj (mgr.14201) 651 : cluster [DBG] pgmap v388: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:28.008 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:06:28.163 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:06:28.163 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (3m) 2m ago 8m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:06:28.163 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 2m ago 8m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:06:28.163 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 112s ago 8m - - 2026-03-07T08:06:28.164 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (8m) 112s ago 8m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:06:28.348 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:06:28.348 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:06:28.348 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:06:29.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:29 vm09 bash[21220]: audit 2026-03-07T07:06:27.994773+0000 mgr.vm07.yrfcuj (mgr.14201) 652 : audit [DBG] from='client.15748 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:29.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:29 vm09 bash[21220]: audit 2026-03-07T07:06:27.994773+0000 mgr.vm07.yrfcuj (mgr.14201) 652 : audit [DBG] from='client.15748 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:29.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:29 vm09 bash[21220]: audit 2026-03-07T07:06:28.162670+0000 mgr.vm07.yrfcuj (mgr.14201) 653 : audit [DBG] from='client.15752 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:29.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:29 vm09 bash[21220]: audit 2026-03-07T07:06:28.162670+0000 mgr.vm07.yrfcuj (mgr.14201) 653 : audit [DBG] from='client.15752 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:29.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:29 vm09 bash[21220]: cluster 2026-03-07T07:06:28.224763+0000 mgr.vm07.yrfcuj (mgr.14201) 654 : cluster [DBG] pgmap v389: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:29.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:29 vm09 bash[21220]: cluster 2026-03-07T07:06:28.224763+0000 mgr.vm07.yrfcuj (mgr.14201) 654 : cluster [DBG] pgmap v389: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:29.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:29 vm09 bash[21220]: audit 2026-03-07T07:06:28.350753+0000 mon.vm07 (mon.0) 1045 : audit [DBG] from='client.? 192.168.123.107:0/1852360541' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:29.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:29 vm09 bash[21220]: audit 2026-03-07T07:06:28.350753+0000 mon.vm07 (mon.0) 1045 : audit [DBG] from='client.? 192.168.123.107:0/1852360541' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:29.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:29 vm07 bash[17031]: audit 2026-03-07T07:06:27.994773+0000 mgr.vm07.yrfcuj (mgr.14201) 652 : audit [DBG] from='client.15748 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:29.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:29 vm07 bash[17031]: audit 2026-03-07T07:06:27.994773+0000 mgr.vm07.yrfcuj (mgr.14201) 652 : audit [DBG] from='client.15748 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:29.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:29 vm07 bash[17031]: audit 2026-03-07T07:06:28.162670+0000 mgr.vm07.yrfcuj (mgr.14201) 653 : audit [DBG] from='client.15752 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:29.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:29 vm07 bash[17031]: audit 2026-03-07T07:06:28.162670+0000 mgr.vm07.yrfcuj (mgr.14201) 653 : audit [DBG] from='client.15752 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:29.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:29 vm07 bash[17031]: cluster 2026-03-07T07:06:28.224763+0000 mgr.vm07.yrfcuj (mgr.14201) 654 : cluster [DBG] pgmap v389: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:29.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:29 vm07 bash[17031]: cluster 2026-03-07T07:06:28.224763+0000 mgr.vm07.yrfcuj (mgr.14201) 654 : cluster [DBG] pgmap v389: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:29.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:29 vm07 bash[17031]: audit 2026-03-07T07:06:28.350753+0000 mon.vm07 (mon.0) 1045 : audit [DBG] from='client.? 192.168.123.107:0/1852360541' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:29.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:29 vm07 bash[17031]: audit 2026-03-07T07:06:28.350753+0000 mon.vm07 (mon.0) 1045 : audit [DBG] from='client.? 192.168.123.107:0/1852360541' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:31.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:31 vm09 bash[21220]: cluster 2026-03-07T07:06:30.225112+0000 mgr.vm07.yrfcuj (mgr.14201) 655 : cluster [DBG] pgmap v390: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:06:31.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:31 vm09 bash[21220]: cluster 2026-03-07T07:06:30.225112+0000 mgr.vm07.yrfcuj (mgr.14201) 655 : cluster [DBG] pgmap v390: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:06:31.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:31 vm07 bash[17031]: cluster 2026-03-07T07:06:30.225112+0000 mgr.vm07.yrfcuj (mgr.14201) 655 : cluster [DBG] pgmap v390: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:06:31.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:31 vm07 bash[17031]: cluster 2026-03-07T07:06:30.225112+0000 mgr.vm07.yrfcuj (mgr.14201) 655 : cluster [DBG] pgmap v390: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:06:33.525 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:06:33.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:33 vm09 bash[21220]: cluster 2026-03-07T07:06:32.225522+0000 mgr.vm07.yrfcuj (mgr.14201) 656 : cluster [DBG] pgmap v391: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:33.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:33 vm09 bash[21220]: cluster 2026-03-07T07:06:32.225522+0000 mgr.vm07.yrfcuj (mgr.14201) 656 : cluster [DBG] pgmap v391: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:33.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:33 vm07 bash[17031]: cluster 2026-03-07T07:06:32.225522+0000 mgr.vm07.yrfcuj (mgr.14201) 656 : cluster [DBG] pgmap v391: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:33.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:33 vm07 bash[17031]: cluster 2026-03-07T07:06:32.225522+0000 mgr.vm07.yrfcuj (mgr.14201) 656 : cluster [DBG] pgmap v391: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:33.671 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:06:33.671 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (3m) 2m ago 8m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:06:33.671 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 2m ago 8m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:06:33.671 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 117s ago 8m - - 2026-03-07T08:06:33.671 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (8m) 117s ago 8m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:06:33.852 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:06:33.852 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:06:33.852 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:06:34.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:34 vm07 bash[17031]: audit 2026-03-07T07:06:33.512999+0000 mgr.vm07.yrfcuj (mgr.14201) 657 : audit [DBG] from='client.15760 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:34.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:34 vm07 bash[17031]: audit 2026-03-07T07:06:33.512999+0000 mgr.vm07.yrfcuj (mgr.14201) 657 : audit [DBG] from='client.15760 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:34.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:34 vm07 bash[17031]: audit 2026-03-07T07:06:33.854978+0000 mon.vm07 (mon.0) 1046 : audit [DBG] from='client.? 192.168.123.107:0/3386577462' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:34.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:34 vm07 bash[17031]: audit 2026-03-07T07:06:33.854978+0000 mon.vm07 (mon.0) 1046 : audit [DBG] from='client.? 192.168.123.107:0/3386577462' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:34.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:34 vm09 bash[21220]: audit 2026-03-07T07:06:33.512999+0000 mgr.vm07.yrfcuj (mgr.14201) 657 : audit [DBG] from='client.15760 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:34.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:34 vm09 bash[21220]: audit 2026-03-07T07:06:33.512999+0000 mgr.vm07.yrfcuj (mgr.14201) 657 : audit [DBG] from='client.15760 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:34.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:34 vm09 bash[21220]: audit 2026-03-07T07:06:33.854978+0000 mon.vm07 (mon.0) 1046 : audit [DBG] from='client.? 192.168.123.107:0/3386577462' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:34.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:34 vm09 bash[21220]: audit 2026-03-07T07:06:33.854978+0000 mon.vm07 (mon.0) 1046 : audit [DBG] from='client.? 192.168.123.107:0/3386577462' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:35.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:35 vm07 bash[17031]: audit 2026-03-07T07:06:33.671618+0000 mgr.vm07.yrfcuj (mgr.14201) 658 : audit [DBG] from='client.25131 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:35.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:35 vm07 bash[17031]: audit 2026-03-07T07:06:33.671618+0000 mgr.vm07.yrfcuj (mgr.14201) 658 : audit [DBG] from='client.25131 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:35.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:35 vm07 bash[17031]: cluster 2026-03-07T07:06:34.225946+0000 mgr.vm07.yrfcuj (mgr.14201) 659 : cluster [DBG] pgmap v392: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:35.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:35 vm07 bash[17031]: cluster 2026-03-07T07:06:34.225946+0000 mgr.vm07.yrfcuj (mgr.14201) 659 : cluster [DBG] pgmap v392: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:35.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:35 vm09 bash[21220]: audit 2026-03-07T07:06:33.671618+0000 mgr.vm07.yrfcuj (mgr.14201) 658 : audit [DBG] from='client.25131 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:35.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:35 vm09 bash[21220]: audit 2026-03-07T07:06:33.671618+0000 mgr.vm07.yrfcuj (mgr.14201) 658 : audit [DBG] from='client.25131 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:35.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:35 vm09 bash[21220]: cluster 2026-03-07T07:06:34.225946+0000 mgr.vm07.yrfcuj (mgr.14201) 659 : cluster [DBG] pgmap v392: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:35.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:35 vm09 bash[21220]: cluster 2026-03-07T07:06:34.225946+0000 mgr.vm07.yrfcuj (mgr.14201) 659 : cluster [DBG] pgmap v392: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:36.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:36 vm07 bash[17031]: audit 2026-03-07T07:06:36.263701+0000 mon.vm07 (mon.0) 1047 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:06:36.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:36 vm07 bash[17031]: audit 2026-03-07T07:06:36.263701+0000 mon.vm07 (mon.0) 1047 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:06:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:36 vm09 bash[21220]: audit 2026-03-07T07:06:36.263701+0000 mon.vm07 (mon.0) 1047 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:06:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:36 vm09 bash[21220]: audit 2026-03-07T07:06:36.263701+0000 mon.vm07 (mon.0) 1047 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:06:37.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:37 vm07 bash[17031]: cluster 2026-03-07T07:06:36.226283+0000 mgr.vm07.yrfcuj (mgr.14201) 660 : cluster [DBG] pgmap v393: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:37.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:37 vm07 bash[17031]: cluster 2026-03-07T07:06:36.226283+0000 mgr.vm07.yrfcuj (mgr.14201) 660 : cluster [DBG] pgmap v393: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:37.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:37 vm07 bash[17031]: audit 2026-03-07T07:06:36.563823+0000 mon.vm07 (mon.0) 1048 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"}]: dispatch 2026-03-07T08:06:37.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:37 vm07 bash[17031]: audit 2026-03-07T07:06:36.563823+0000 mon.vm07 (mon.0) 1048 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"}]: dispatch 2026-03-07T08:06:37.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:37 vm07 bash[17031]: audit 2026-03-07T07:06:36.564362+0000 mon.vm07 (mon.0) 1049 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-07T08:06:37.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:37 vm07 bash[17031]: audit 2026-03-07T07:06:36.564362+0000 mon.vm07 (mon.0) 1049 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-07T08:06:37.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:37 vm07 bash[17031]: audit 2026-03-07T07:06:36.565564+0000 mon.vm07 (mon.0) 1050 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:06:37.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:37 vm07 bash[17031]: audit 2026-03-07T07:06:36.565564+0000 mon.vm07 (mon.0) 1050 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:06:37.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:37 vm07 bash[17031]: audit 2026-03-07T07:06:36.566315+0000 mon.vm07 (mon.0) 1051 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:06:37.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:37 vm07 bash[17031]: audit 2026-03-07T07:06:36.566315+0000 mon.vm07 (mon.0) 1051 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:06:37.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:37 vm07 bash[17031]: cluster 2026-03-07T07:06:36.567539+0000 mgr.vm07.yrfcuj (mgr.14201) 661 : cluster [DBG] pgmap v394: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:37.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:37 vm07 bash[17031]: cluster 2026-03-07T07:06:36.567539+0000 mgr.vm07.yrfcuj (mgr.14201) 661 : cluster [DBG] pgmap v394: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:37.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:37 vm07 bash[17031]: audit 2026-03-07T07:06:36.572092+0000 mon.vm07 (mon.0) 1052 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:06:37.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:37 vm07 bash[17031]: audit 2026-03-07T07:06:36.572092+0000 mon.vm07 (mon.0) 1052 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:06:37.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:37 vm07 bash[17031]: audit 2026-03-07T07:06:36.574266+0000 mon.vm07 (mon.0) 1053 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:06:37.651 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:37 vm07 bash[17031]: audit 2026-03-07T07:06:36.574266+0000 mon.vm07 (mon.0) 1053 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:06:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:37 vm09 bash[21220]: cluster 2026-03-07T07:06:36.226283+0000 mgr.vm07.yrfcuj (mgr.14201) 660 : cluster [DBG] pgmap v393: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:37 vm09 bash[21220]: cluster 2026-03-07T07:06:36.226283+0000 mgr.vm07.yrfcuj (mgr.14201) 660 : cluster [DBG] pgmap v393: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:37 vm09 bash[21220]: audit 2026-03-07T07:06:36.563823+0000 mon.vm07 (mon.0) 1048 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"}]: dispatch 2026-03-07T08:06:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:37 vm09 bash[21220]: audit 2026-03-07T07:06:36.563823+0000 mon.vm07 (mon.0) 1048 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"}]: dispatch 2026-03-07T08:06:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:37 vm09 bash[21220]: audit 2026-03-07T07:06:36.564362+0000 mon.vm07 (mon.0) 1049 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-07T08:06:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:37 vm09 bash[21220]: audit 2026-03-07T07:06:36.564362+0000 mon.vm07 (mon.0) 1049 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-07T08:06:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:37 vm09 bash[21220]: audit 2026-03-07T07:06:36.565564+0000 mon.vm07 (mon.0) 1050 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:06:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:37 vm09 bash[21220]: audit 2026-03-07T07:06:36.565564+0000 mon.vm07 (mon.0) 1050 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:06:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:37 vm09 bash[21220]: audit 2026-03-07T07:06:36.566315+0000 mon.vm07 (mon.0) 1051 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:06:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:37 vm09 bash[21220]: audit 2026-03-07T07:06:36.566315+0000 mon.vm07 (mon.0) 1051 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:06:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:37 vm09 bash[21220]: cluster 2026-03-07T07:06:36.567539+0000 mgr.vm07.yrfcuj (mgr.14201) 661 : cluster [DBG] pgmap v394: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:37 vm09 bash[21220]: cluster 2026-03-07T07:06:36.567539+0000 mgr.vm07.yrfcuj (mgr.14201) 661 : cluster [DBG] pgmap v394: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:37 vm09 bash[21220]: audit 2026-03-07T07:06:36.572092+0000 mon.vm07 (mon.0) 1052 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:06:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:37 vm09 bash[21220]: audit 2026-03-07T07:06:36.572092+0000 mon.vm07 (mon.0) 1052 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:06:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:37 vm09 bash[21220]: audit 2026-03-07T07:06:36.574266+0000 mon.vm07 (mon.0) 1053 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:06:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:37 vm09 bash[21220]: audit 2026-03-07T07:06:36.574266+0000 mon.vm07 (mon.0) 1053 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:06:38.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:38 vm07 bash[17031]: cluster 2026-03-07T07:06:38.568057+0000 mgr.vm07.yrfcuj (mgr.14201) 662 : cluster [DBG] pgmap v395: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:38.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:38 vm07 bash[17031]: cluster 2026-03-07T07:06:38.568057+0000 mgr.vm07.yrfcuj (mgr.14201) 662 : cluster [DBG] pgmap v395: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:39.020 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:06:39.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:38 vm09 bash[21220]: cluster 2026-03-07T07:06:38.568057+0000 mgr.vm07.yrfcuj (mgr.14201) 662 : cluster [DBG] pgmap v395: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:39.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:38 vm09 bash[21220]: cluster 2026-03-07T07:06:38.568057+0000 mgr.vm07.yrfcuj (mgr.14201) 662 : cluster [DBG] pgmap v395: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:39.169 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:06:39.169 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (3m) 2m ago 9m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:06:39.169 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 2m ago 8m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:06:39.169 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 2m ago 9m - - 2026-03-07T08:06:39.169 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (9m) 2m ago 9m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:06:39.358 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:06:39.358 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:06:39.358 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:06:39.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:39 vm07 bash[17031]: audit 2026-03-07T07:06:39.008681+0000 mgr.vm07.yrfcuj (mgr.14201) 663 : audit [DBG] from='client.15772 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:39.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:39 vm07 bash[17031]: audit 2026-03-07T07:06:39.008681+0000 mgr.vm07.yrfcuj (mgr.14201) 663 : audit [DBG] from='client.15772 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:39.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:39 vm07 bash[17031]: audit 2026-03-07T07:06:39.168419+0000 mgr.vm07.yrfcuj (mgr.14201) 664 : audit [DBG] from='client.15776 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:39.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:39 vm07 bash[17031]: audit 2026-03-07T07:06:39.168419+0000 mgr.vm07.yrfcuj (mgr.14201) 664 : audit [DBG] from='client.15776 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:39.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:39 vm07 bash[17031]: audit 2026-03-07T07:06:39.360196+0000 mon.vm07 (mon.0) 1054 : audit [DBG] from='client.? 192.168.123.107:0/139794155' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:39.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:39 vm07 bash[17031]: audit 2026-03-07T07:06:39.360196+0000 mon.vm07 (mon.0) 1054 : audit [DBG] from='client.? 192.168.123.107:0/139794155' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:39 vm09 bash[21220]: audit 2026-03-07T07:06:39.008681+0000 mgr.vm07.yrfcuj (mgr.14201) 663 : audit [DBG] from='client.15772 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:39 vm09 bash[21220]: audit 2026-03-07T07:06:39.008681+0000 mgr.vm07.yrfcuj (mgr.14201) 663 : audit [DBG] from='client.15772 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:39 vm09 bash[21220]: audit 2026-03-07T07:06:39.168419+0000 mgr.vm07.yrfcuj (mgr.14201) 664 : audit [DBG] from='client.15776 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:39 vm09 bash[21220]: audit 2026-03-07T07:06:39.168419+0000 mgr.vm07.yrfcuj (mgr.14201) 664 : audit [DBG] from='client.15776 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:39 vm09 bash[21220]: audit 2026-03-07T07:06:39.360196+0000 mon.vm07 (mon.0) 1054 : audit [DBG] from='client.? 192.168.123.107:0/139794155' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:39 vm09 bash[21220]: audit 2026-03-07T07:06:39.360196+0000 mon.vm07 (mon.0) 1054 : audit [DBG] from='client.? 192.168.123.107:0/139794155' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:40.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:40 vm07 bash[17031]: audit 2026-03-07T07:06:40.137684+0000 mon.vm07 (mon.0) 1055 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:06:40.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:40 vm07 bash[17031]: audit 2026-03-07T07:06:40.137684+0000 mon.vm07 (mon.0) 1055 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:06:40.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:40 vm07 bash[17031]: cluster 2026-03-07T07:06:40.568434+0000 mgr.vm07.yrfcuj (mgr.14201) 665 : cluster [DBG] pgmap v396: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:40.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:40 vm07 bash[17031]: cluster 2026-03-07T07:06:40.568434+0000 mgr.vm07.yrfcuj (mgr.14201) 665 : cluster [DBG] pgmap v396: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:41.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:40 vm09 bash[21220]: audit 2026-03-07T07:06:40.137684+0000 mon.vm07 (mon.0) 1055 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:06:41.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:40 vm09 bash[21220]: audit 2026-03-07T07:06:40.137684+0000 mon.vm07 (mon.0) 1055 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:06:41.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:40 vm09 bash[21220]: cluster 2026-03-07T07:06:40.568434+0000 mgr.vm07.yrfcuj (mgr.14201) 665 : cluster [DBG] pgmap v396: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:41.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:40 vm09 bash[21220]: cluster 2026-03-07T07:06:40.568434+0000 mgr.vm07.yrfcuj (mgr.14201) 665 : cluster [DBG] pgmap v396: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:42.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:42 vm07 bash[17031]: cluster 2026-03-07T07:06:42.568820+0000 mgr.vm07.yrfcuj (mgr.14201) 666 : cluster [DBG] pgmap v397: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:42.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:42 vm07 bash[17031]: cluster 2026-03-07T07:06:42.568820+0000 mgr.vm07.yrfcuj (mgr.14201) 666 : cluster [DBG] pgmap v397: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:43.091 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:42 vm09 bash[21220]: cluster 2026-03-07T07:06:42.568820+0000 mgr.vm07.yrfcuj (mgr.14201) 666 : cluster [DBG] pgmap v397: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:43.091 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:42 vm09 bash[21220]: cluster 2026-03-07T07:06:42.568820+0000 mgr.vm07.yrfcuj (mgr.14201) 666 : cluster [DBG] pgmap v397: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:44.521 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:06:44.670 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:06:44.670 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (3m) 2m ago 9m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:06:44.670 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 2m ago 9m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:06:44.670 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 2m ago 9m - - 2026-03-07T08:06:44.670 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (9m) 2m ago 9m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:06:44.845 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:06:44.845 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:06:44.845 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:06:44.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:44 vm07 bash[17031]: audit 2026-03-07T07:06:44.510665+0000 mgr.vm07.yrfcuj (mgr.14201) 667 : audit [DBG] from='client.15784 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:44.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:44 vm07 bash[17031]: audit 2026-03-07T07:06:44.510665+0000 mgr.vm07.yrfcuj (mgr.14201) 667 : audit [DBG] from='client.15784 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:44.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:44 vm07 bash[17031]: cluster 2026-03-07T07:06:44.569229+0000 mgr.vm07.yrfcuj (mgr.14201) 668 : cluster [DBG] pgmap v398: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:44.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:44 vm07 bash[17031]: cluster 2026-03-07T07:06:44.569229+0000 mgr.vm07.yrfcuj (mgr.14201) 668 : cluster [DBG] pgmap v398: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:45.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:44 vm09 bash[21220]: audit 2026-03-07T07:06:44.510665+0000 mgr.vm07.yrfcuj (mgr.14201) 667 : audit [DBG] from='client.15784 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:45.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:44 vm09 bash[21220]: audit 2026-03-07T07:06:44.510665+0000 mgr.vm07.yrfcuj (mgr.14201) 667 : audit [DBG] from='client.15784 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:45.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:44 vm09 bash[21220]: cluster 2026-03-07T07:06:44.569229+0000 mgr.vm07.yrfcuj (mgr.14201) 668 : cluster [DBG] pgmap v398: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:45.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:44 vm09 bash[21220]: cluster 2026-03-07T07:06:44.569229+0000 mgr.vm07.yrfcuj (mgr.14201) 668 : cluster [DBG] pgmap v398: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:45.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:45 vm07 bash[17031]: audit 2026-03-07T07:06:44.669995+0000 mgr.vm07.yrfcuj (mgr.14201) 669 : audit [DBG] from='client.25147 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:45.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:45 vm07 bash[17031]: audit 2026-03-07T07:06:44.669995+0000 mgr.vm07.yrfcuj (mgr.14201) 669 : audit [DBG] from='client.25147 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:45.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:45 vm07 bash[17031]: audit 2026-03-07T07:06:44.847331+0000 mon.vm07 (mon.0) 1056 : audit [DBG] from='client.? 192.168.123.107:0/2997138378' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:45.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:45 vm07 bash[17031]: audit 2026-03-07T07:06:44.847331+0000 mon.vm07 (mon.0) 1056 : audit [DBG] from='client.? 192.168.123.107:0/2997138378' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:46.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:45 vm09 bash[21220]: audit 2026-03-07T07:06:44.669995+0000 mgr.vm07.yrfcuj (mgr.14201) 669 : audit [DBG] from='client.25147 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:46.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:45 vm09 bash[21220]: audit 2026-03-07T07:06:44.669995+0000 mgr.vm07.yrfcuj (mgr.14201) 669 : audit [DBG] from='client.25147 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:46.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:45 vm09 bash[21220]: audit 2026-03-07T07:06:44.847331+0000 mon.vm07 (mon.0) 1056 : audit [DBG] from='client.? 192.168.123.107:0/2997138378' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:46.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:45 vm09 bash[21220]: audit 2026-03-07T07:06:44.847331+0000 mon.vm07 (mon.0) 1056 : audit [DBG] from='client.? 192.168.123.107:0/2997138378' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:46.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:46 vm07 bash[17031]: cluster 2026-03-07T07:06:46.569564+0000 mgr.vm07.yrfcuj (mgr.14201) 670 : cluster [DBG] pgmap v399: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:46.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:46 vm07 bash[17031]: cluster 2026-03-07T07:06:46.569564+0000 mgr.vm07.yrfcuj (mgr.14201) 670 : cluster [DBG] pgmap v399: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:47.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:46 vm09 bash[21220]: cluster 2026-03-07T07:06:46.569564+0000 mgr.vm07.yrfcuj (mgr.14201) 670 : cluster [DBG] pgmap v399: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:47.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:46 vm09 bash[21220]: cluster 2026-03-07T07:06:46.569564+0000 mgr.vm07.yrfcuj (mgr.14201) 670 : cluster [DBG] pgmap v399: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:06:48.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:48 vm07 bash[17031]: cluster 2026-03-07T07:06:48.569967+0000 mgr.vm07.yrfcuj (mgr.14201) 671 : cluster [DBG] pgmap v400: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:48.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:48 vm07 bash[17031]: cluster 2026-03-07T07:06:48.569967+0000 mgr.vm07.yrfcuj (mgr.14201) 671 : cluster [DBG] pgmap v400: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:49.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:48 vm09 bash[21220]: cluster 2026-03-07T07:06:48.569967+0000 mgr.vm07.yrfcuj (mgr.14201) 671 : cluster [DBG] pgmap v400: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:49.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:48 vm09 bash[21220]: cluster 2026-03-07T07:06:48.569967+0000 mgr.vm07.yrfcuj (mgr.14201) 671 : cluster [DBG] pgmap v400: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:50.021 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:06:50.180 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:06:50.180 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (3m) 2m ago 9m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:06:50.180 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 2m ago 9m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:06:50.180 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 2m ago 9m - - 2026-03-07T08:06:50.180 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (9m) 2m ago 9m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:06:50.365 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:06:50.366 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:06:50.366 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:06:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:50 vm09 bash[21220]: audit 2026-03-07T07:06:50.364425+0000 mon.vm09 (mon.1) 42 : audit [DBG] from='client.? 192.168.123.107:0/1701977231' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:50 vm09 bash[21220]: audit 2026-03-07T07:06:50.364425+0000 mon.vm09 (mon.1) 42 : audit [DBG] from='client.? 192.168.123.107:0/1701977231' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:50.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:50 vm07 bash[17031]: audit 2026-03-07T07:06:50.364425+0000 mon.vm09 (mon.1) 42 : audit [DBG] from='client.? 192.168.123.107:0/1701977231' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:50.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:50 vm07 bash[17031]: audit 2026-03-07T07:06:50.364425+0000 mon.vm09 (mon.1) 42 : audit [DBG] from='client.? 192.168.123.107:0/1701977231' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:51 vm09 bash[21220]: audit 2026-03-07T07:06:50.009115+0000 mgr.vm07.yrfcuj (mgr.14201) 672 : audit [DBG] from='client.15796 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:51 vm09 bash[21220]: audit 2026-03-07T07:06:50.009115+0000 mgr.vm07.yrfcuj (mgr.14201) 672 : audit [DBG] from='client.15796 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:51 vm09 bash[21220]: audit 2026-03-07T07:06:50.180045+0000 mgr.vm07.yrfcuj (mgr.14201) 673 : audit [DBG] from='client.15800 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:51 vm09 bash[21220]: audit 2026-03-07T07:06:50.180045+0000 mgr.vm07.yrfcuj (mgr.14201) 673 : audit [DBG] from='client.15800 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:51 vm09 bash[21220]: cluster 2026-03-07T07:06:50.570368+0000 mgr.vm07.yrfcuj (mgr.14201) 674 : cluster [DBG] pgmap v401: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:51 vm09 bash[21220]: cluster 2026-03-07T07:06:50.570368+0000 mgr.vm07.yrfcuj (mgr.14201) 674 : cluster [DBG] pgmap v401: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:51.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:51 vm07 bash[17031]: audit 2026-03-07T07:06:50.009115+0000 mgr.vm07.yrfcuj (mgr.14201) 672 : audit [DBG] from='client.15796 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:51.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:51 vm07 bash[17031]: audit 2026-03-07T07:06:50.009115+0000 mgr.vm07.yrfcuj (mgr.14201) 672 : audit [DBG] from='client.15796 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:51.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:51 vm07 bash[17031]: audit 2026-03-07T07:06:50.180045+0000 mgr.vm07.yrfcuj (mgr.14201) 673 : audit [DBG] from='client.15800 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:51.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:51 vm07 bash[17031]: audit 2026-03-07T07:06:50.180045+0000 mgr.vm07.yrfcuj (mgr.14201) 673 : audit [DBG] from='client.15800 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:51.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:51 vm07 bash[17031]: cluster 2026-03-07T07:06:50.570368+0000 mgr.vm07.yrfcuj (mgr.14201) 674 : cluster [DBG] pgmap v401: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:51.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:51 vm07 bash[17031]: cluster 2026-03-07T07:06:50.570368+0000 mgr.vm07.yrfcuj (mgr.14201) 674 : cluster [DBG] pgmap v401: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:52.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:52 vm07 bash[17031]: cluster 2026-03-07T07:06:52.570797+0000 mgr.vm07.yrfcuj (mgr.14201) 675 : cluster [DBG] pgmap v402: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:52.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:52 vm07 bash[17031]: cluster 2026-03-07T07:06:52.570797+0000 mgr.vm07.yrfcuj (mgr.14201) 675 : cluster [DBG] pgmap v402: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:53.091 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:52 vm09 bash[21220]: cluster 2026-03-07T07:06:52.570797+0000 mgr.vm07.yrfcuj (mgr.14201) 675 : cluster [DBG] pgmap v402: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:53.091 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:52 vm09 bash[21220]: cluster 2026-03-07T07:06:52.570797+0000 mgr.vm07.yrfcuj (mgr.14201) 675 : cluster [DBG] pgmap v402: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:06:54.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:54 vm07 bash[17031]: cluster 2026-03-07T07:06:54.571335+0000 mgr.vm07.yrfcuj (mgr.14201) 676 : cluster [DBG] pgmap v403: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:54.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:54 vm07 bash[17031]: cluster 2026-03-07T07:06:54.571335+0000 mgr.vm07.yrfcuj (mgr.14201) 676 : cluster [DBG] pgmap v403: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:55.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:54 vm09 bash[21220]: cluster 2026-03-07T07:06:54.571335+0000 mgr.vm07.yrfcuj (mgr.14201) 676 : cluster [DBG] pgmap v403: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:55.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:54 vm09 bash[21220]: cluster 2026-03-07T07:06:54.571335+0000 mgr.vm07.yrfcuj (mgr.14201) 676 : cluster [DBG] pgmap v403: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:55.540 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:06:55.692 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:06:55.692 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (3m) 2m ago 9m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:06:55.692 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 2m ago 9m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:06:55.692 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 2m ago 9m - - 2026-03-07T08:06:55.692 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (9m) 2m ago 9m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:06:55.881 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:06:55.881 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:06:55.881 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:06:55.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:55 vm07 bash[17031]: audit 2026-03-07T07:06:55.137989+0000 mon.vm07 (mon.0) 1057 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:06:55.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:55 vm07 bash[17031]: audit 2026-03-07T07:06:55.137989+0000 mon.vm07 (mon.0) 1057 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:06:55.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:55 vm07 bash[17031]: audit 2026-03-07T07:06:55.526725+0000 mgr.vm07.yrfcuj (mgr.14201) 677 : audit [DBG] from='client.15808 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:55.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:55 vm07 bash[17031]: audit 2026-03-07T07:06:55.526725+0000 mgr.vm07.yrfcuj (mgr.14201) 677 : audit [DBG] from='client.15808 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:56.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:55 vm09 bash[21220]: audit 2026-03-07T07:06:55.137989+0000 mon.vm07 (mon.0) 1057 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:06:56.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:55 vm09 bash[21220]: audit 2026-03-07T07:06:55.137989+0000 mon.vm07 (mon.0) 1057 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:06:56.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:55 vm09 bash[21220]: audit 2026-03-07T07:06:55.526725+0000 mgr.vm07.yrfcuj (mgr.14201) 677 : audit [DBG] from='client.15808 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:56.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:55 vm09 bash[21220]: audit 2026-03-07T07:06:55.526725+0000 mgr.vm07.yrfcuj (mgr.14201) 677 : audit [DBG] from='client.15808 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:56.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:56 vm07 bash[17031]: audit 2026-03-07T07:06:55.691567+0000 mgr.vm07.yrfcuj (mgr.14201) 678 : audit [DBG] from='client.15812 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:56.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:56 vm07 bash[17031]: audit 2026-03-07T07:06:55.691567+0000 mgr.vm07.yrfcuj (mgr.14201) 678 : audit [DBG] from='client.15812 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:56.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:56 vm07 bash[17031]: audit 2026-03-07T07:06:55.880343+0000 mon.vm09 (mon.1) 43 : audit [DBG] from='client.? 192.168.123.107:0/1041354724' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:56.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:56 vm07 bash[17031]: audit 2026-03-07T07:06:55.880343+0000 mon.vm09 (mon.1) 43 : audit [DBG] from='client.? 192.168.123.107:0/1041354724' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:56.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:56 vm07 bash[17031]: cluster 2026-03-07T07:06:56.571766+0000 mgr.vm07.yrfcuj (mgr.14201) 679 : cluster [DBG] pgmap v404: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:56.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:56 vm07 bash[17031]: cluster 2026-03-07T07:06:56.571766+0000 mgr.vm07.yrfcuj (mgr.14201) 679 : cluster [DBG] pgmap v404: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:57.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:56 vm09 bash[21220]: audit 2026-03-07T07:06:55.691567+0000 mgr.vm07.yrfcuj (mgr.14201) 678 : audit [DBG] from='client.15812 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:57.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:56 vm09 bash[21220]: audit 2026-03-07T07:06:55.691567+0000 mgr.vm07.yrfcuj (mgr.14201) 678 : audit [DBG] from='client.15812 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:06:57.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:56 vm09 bash[21220]: audit 2026-03-07T07:06:55.880343+0000 mon.vm09 (mon.1) 43 : audit [DBG] from='client.? 192.168.123.107:0/1041354724' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:57.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:56 vm09 bash[21220]: audit 2026-03-07T07:06:55.880343+0000 mon.vm09 (mon.1) 43 : audit [DBG] from='client.? 192.168.123.107:0/1041354724' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:06:57.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:56 vm09 bash[21220]: cluster 2026-03-07T07:06:56.571766+0000 mgr.vm07.yrfcuj (mgr.14201) 679 : cluster [DBG] pgmap v404: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:57.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:56 vm09 bash[21220]: cluster 2026-03-07T07:06:56.571766+0000 mgr.vm07.yrfcuj (mgr.14201) 679 : cluster [DBG] pgmap v404: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:58.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:58 vm07 bash[17031]: cluster 2026-03-07T07:06:58.572358+0000 mgr.vm07.yrfcuj (mgr.14201) 680 : cluster [DBG] pgmap v405: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:58.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:06:58 vm07 bash[17031]: cluster 2026-03-07T07:06:58.572358+0000 mgr.vm07.yrfcuj (mgr.14201) 680 : cluster [DBG] pgmap v405: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:59.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:58 vm09 bash[21220]: cluster 2026-03-07T07:06:58.572358+0000 mgr.vm07.yrfcuj (mgr.14201) 680 : cluster [DBG] pgmap v405: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:06:59.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:06:58 vm09 bash[21220]: cluster 2026-03-07T07:06:58.572358+0000 mgr.vm07.yrfcuj (mgr.14201) 680 : cluster [DBG] pgmap v405: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:00.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:00 vm07 bash[17031]: cluster 2026-03-07T07:07:00.572708+0000 mgr.vm07.yrfcuj (mgr.14201) 681 : cluster [DBG] pgmap v406: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:07:00.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:00 vm07 bash[17031]: cluster 2026-03-07T07:07:00.572708+0000 mgr.vm07.yrfcuj (mgr.14201) 681 : cluster [DBG] pgmap v406: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:07:01.057 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:07:01.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:00 vm09 bash[21220]: cluster 2026-03-07T07:07:00.572708+0000 mgr.vm07.yrfcuj (mgr.14201) 681 : cluster [DBG] pgmap v406: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:07:01.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:00 vm09 bash[21220]: cluster 2026-03-07T07:07:00.572708+0000 mgr.vm07.yrfcuj (mgr.14201) 681 : cluster [DBG] pgmap v406: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:07:01.229 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:07:01.229 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (3m) 2m ago 9m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:07:01.230 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 2m ago 9m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:07:01.230 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 2m ago 9m - - 2026-03-07T08:07:01.230 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (9m) 2m ago 9m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:07:01.422 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:07:01.422 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:07:01.422 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:07:01.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:01 vm07 bash[17031]: audit 2026-03-07T07:07:01.045926+0000 mgr.vm07.yrfcuj (mgr.14201) 682 : audit [DBG] from='client.15820 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:01.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:01 vm07 bash[17031]: audit 2026-03-07T07:07:01.045926+0000 mgr.vm07.yrfcuj (mgr.14201) 682 : audit [DBG] from='client.15820 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:01.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:01 vm07 bash[17031]: audit 2026-03-07T07:07:01.229387+0000 mgr.vm07.yrfcuj (mgr.14201) 683 : audit [DBG] from='client.15824 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:01.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:01 vm07 bash[17031]: audit 2026-03-07T07:07:01.229387+0000 mgr.vm07.yrfcuj (mgr.14201) 683 : audit [DBG] from='client.15824 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:01.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:01 vm07 bash[17031]: audit 2026-03-07T07:07:01.421800+0000 mon.vm09 (mon.1) 44 : audit [DBG] from='client.? 192.168.123.107:0/693933003' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:01.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:01 vm07 bash[17031]: audit 2026-03-07T07:07:01.421800+0000 mon.vm09 (mon.1) 44 : audit [DBG] from='client.? 192.168.123.107:0/693933003' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:02.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:01 vm09 bash[21220]: audit 2026-03-07T07:07:01.045926+0000 mgr.vm07.yrfcuj (mgr.14201) 682 : audit [DBG] from='client.15820 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:02.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:01 vm09 bash[21220]: audit 2026-03-07T07:07:01.045926+0000 mgr.vm07.yrfcuj (mgr.14201) 682 : audit [DBG] from='client.15820 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:02.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:01 vm09 bash[21220]: audit 2026-03-07T07:07:01.229387+0000 mgr.vm07.yrfcuj (mgr.14201) 683 : audit [DBG] from='client.15824 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:02.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:01 vm09 bash[21220]: audit 2026-03-07T07:07:01.229387+0000 mgr.vm07.yrfcuj (mgr.14201) 683 : audit [DBG] from='client.15824 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:02.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:01 vm09 bash[21220]: audit 2026-03-07T07:07:01.421800+0000 mon.vm09 (mon.1) 44 : audit [DBG] from='client.? 192.168.123.107:0/693933003' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:02.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:01 vm09 bash[21220]: audit 2026-03-07T07:07:01.421800+0000 mon.vm09 (mon.1) 44 : audit [DBG] from='client.? 192.168.123.107:0/693933003' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:02.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:02 vm07 bash[17031]: cluster 2026-03-07T07:07:02.573121+0000 mgr.vm07.yrfcuj (mgr.14201) 684 : cluster [DBG] pgmap v407: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:02.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:02 vm07 bash[17031]: cluster 2026-03-07T07:07:02.573121+0000 mgr.vm07.yrfcuj (mgr.14201) 684 : cluster [DBG] pgmap v407: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:03.091 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:02 vm09 bash[21220]: cluster 2026-03-07T07:07:02.573121+0000 mgr.vm07.yrfcuj (mgr.14201) 684 : cluster [DBG] pgmap v407: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:03.091 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:02 vm09 bash[21220]: cluster 2026-03-07T07:07:02.573121+0000 mgr.vm07.yrfcuj (mgr.14201) 684 : cluster [DBG] pgmap v407: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:04.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:04 vm07 bash[17031]: cluster 2026-03-07T07:07:04.573554+0000 mgr.vm07.yrfcuj (mgr.14201) 685 : cluster [DBG] pgmap v408: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:04.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:04 vm07 bash[17031]: cluster 2026-03-07T07:07:04.573554+0000 mgr.vm07.yrfcuj (mgr.14201) 685 : cluster [DBG] pgmap v408: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:05.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:04 vm09 bash[21220]: cluster 2026-03-07T07:07:04.573554+0000 mgr.vm07.yrfcuj (mgr.14201) 685 : cluster [DBG] pgmap v408: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:05.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:04 vm09 bash[21220]: cluster 2026-03-07T07:07:04.573554+0000 mgr.vm07.yrfcuj (mgr.14201) 685 : cluster [DBG] pgmap v408: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:06.607 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:07:06.762 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:07:06.762 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (3m) 2m ago 9m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:07:06.762 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (2m) 2m ago 9m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:07:06.762 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 2m ago 9m - - 2026-03-07T08:07:06.762 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (9m) 2m ago 9m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:07:06.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:06 vm07 bash[17031]: cluster 2026-03-07T07:07:06.573878+0000 mgr.vm07.yrfcuj (mgr.14201) 686 : cluster [DBG] pgmap v409: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:06.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:06 vm07 bash[17031]: cluster 2026-03-07T07:07:06.573878+0000 mgr.vm07.yrfcuj (mgr.14201) 686 : cluster [DBG] pgmap v409: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:06.950 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:07:06.951 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:07:06.951 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:07:07.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:06 vm09 bash[21220]: cluster 2026-03-07T07:07:06.573878+0000 mgr.vm07.yrfcuj (mgr.14201) 686 : cluster [DBG] pgmap v409: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:07.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:06 vm09 bash[21220]: cluster 2026-03-07T07:07:06.573878+0000 mgr.vm07.yrfcuj (mgr.14201) 686 : cluster [DBG] pgmap v409: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:07.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:07 vm07 bash[17031]: audit 2026-03-07T07:07:06.593063+0000 mgr.vm07.yrfcuj (mgr.14201) 687 : audit [DBG] from='client.15832 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:07.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:07 vm07 bash[17031]: audit 2026-03-07T07:07:06.593063+0000 mgr.vm07.yrfcuj (mgr.14201) 687 : audit [DBG] from='client.15832 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:07.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:07 vm07 bash[17031]: audit 2026-03-07T07:07:06.762109+0000 mgr.vm07.yrfcuj (mgr.14201) 688 : audit [DBG] from='client.15836 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:07.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:07 vm07 bash[17031]: audit 2026-03-07T07:07:06.762109+0000 mgr.vm07.yrfcuj (mgr.14201) 688 : audit [DBG] from='client.15836 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:07.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:07 vm07 bash[17031]: audit 2026-03-07T07:07:06.953511+0000 mon.vm07 (mon.0) 1058 : audit [DBG] from='client.? 192.168.123.107:0/1543878605' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:07.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:07 vm07 bash[17031]: audit 2026-03-07T07:07:06.953511+0000 mon.vm07 (mon.0) 1058 : audit [DBG] from='client.? 192.168.123.107:0/1543878605' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:08.092 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:07 vm09 bash[21220]: audit 2026-03-07T07:07:06.593063+0000 mgr.vm07.yrfcuj (mgr.14201) 687 : audit [DBG] from='client.15832 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:08.092 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:07 vm09 bash[21220]: audit 2026-03-07T07:07:06.593063+0000 mgr.vm07.yrfcuj (mgr.14201) 687 : audit [DBG] from='client.15832 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:08.092 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:07 vm09 bash[21220]: audit 2026-03-07T07:07:06.762109+0000 mgr.vm07.yrfcuj (mgr.14201) 688 : audit [DBG] from='client.15836 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:08.092 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:07 vm09 bash[21220]: audit 2026-03-07T07:07:06.762109+0000 mgr.vm07.yrfcuj (mgr.14201) 688 : audit [DBG] from='client.15836 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:08.092 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:07 vm09 bash[21220]: audit 2026-03-07T07:07:06.953511+0000 mon.vm07 (mon.0) 1058 : audit [DBG] from='client.? 192.168.123.107:0/1543878605' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:08.092 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:07 vm09 bash[21220]: audit 2026-03-07T07:07:06.953511+0000 mon.vm07 (mon.0) 1058 : audit [DBG] from='client.? 192.168.123.107:0/1543878605' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:08.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:08 vm07 bash[17031]: cluster 2026-03-07T07:07:08.574373+0000 mgr.vm07.yrfcuj (mgr.14201) 689 : cluster [DBG] pgmap v410: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:08.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:08 vm07 bash[17031]: cluster 2026-03-07T07:07:08.574373+0000 mgr.vm07.yrfcuj (mgr.14201) 689 : cluster [DBG] pgmap v410: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:09.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:08 vm09 bash[21220]: cluster 2026-03-07T07:07:08.574373+0000 mgr.vm07.yrfcuj (mgr.14201) 689 : cluster [DBG] pgmap v410: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:09.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:08 vm09 bash[21220]: cluster 2026-03-07T07:07:08.574373+0000 mgr.vm07.yrfcuj (mgr.14201) 689 : cluster [DBG] pgmap v410: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:10 vm09 bash[21220]: audit 2026-03-07T07:07:10.138068+0000 mon.vm07 (mon.0) 1059 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:07:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:10 vm09 bash[21220]: audit 2026-03-07T07:07:10.138068+0000 mon.vm07 (mon.0) 1059 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:07:10.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:10 vm07 bash[17031]: audit 2026-03-07T07:07:10.138068+0000 mon.vm07 (mon.0) 1059 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:07:10.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:10 vm07 bash[17031]: audit 2026-03-07T07:07:10.138068+0000 mon.vm07 (mon.0) 1059 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:07:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:11 vm09 bash[21220]: cluster 2026-03-07T07:07:10.574820+0000 mgr.vm07.yrfcuj (mgr.14201) 690 : cluster [DBG] pgmap v411: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:11 vm09 bash[21220]: cluster 2026-03-07T07:07:10.574820+0000 mgr.vm07.yrfcuj (mgr.14201) 690 : cluster [DBG] pgmap v411: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:11.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:11 vm07 bash[17031]: cluster 2026-03-07T07:07:10.574820+0000 mgr.vm07.yrfcuj (mgr.14201) 690 : cluster [DBG] pgmap v411: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:11.650 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:11 vm07 bash[17031]: cluster 2026-03-07T07:07:10.574820+0000 mgr.vm07.yrfcuj (mgr.14201) 690 : cluster [DBG] pgmap v411: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:12.133 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:07:12.287 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:07:12.287 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (3m) 3m ago 9m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:07:12.287 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 3m ago 9m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:07:12.287 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 2m ago 9m - - 2026-03-07T08:07:12.287 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (9m) 2m ago 9m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:07:12.476 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:07:12.476 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:07:12.476 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:07:12.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:12 vm09 bash[21220]: audit 2026-03-07T07:07:12.479111+0000 mon.vm07 (mon.0) 1060 : audit [DBG] from='client.? 192.168.123.107:0/1484372393' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:12.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:12 vm09 bash[21220]: audit 2026-03-07T07:07:12.479111+0000 mon.vm07 (mon.0) 1060 : audit [DBG] from='client.? 192.168.123.107:0/1484372393' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:12.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:12 vm07 bash[17031]: audit 2026-03-07T07:07:12.479111+0000 mon.vm07 (mon.0) 1060 : audit [DBG] from='client.? 192.168.123.107:0/1484372393' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:12.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:12 vm07 bash[17031]: audit 2026-03-07T07:07:12.479111+0000 mon.vm07 (mon.0) 1060 : audit [DBG] from='client.? 192.168.123.107:0/1484372393' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:13.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:13 vm09 bash[21220]: audit 2026-03-07T07:07:12.120464+0000 mgr.vm07.yrfcuj (mgr.14201) 691 : audit [DBG] from='client.15844 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:13.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:13 vm09 bash[21220]: audit 2026-03-07T07:07:12.120464+0000 mgr.vm07.yrfcuj (mgr.14201) 691 : audit [DBG] from='client.15844 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:13.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:13 vm09 bash[21220]: audit 2026-03-07T07:07:12.287734+0000 mgr.vm07.yrfcuj (mgr.14201) 692 : audit [DBG] from='client.15848 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:13.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:13 vm09 bash[21220]: audit 2026-03-07T07:07:12.287734+0000 mgr.vm07.yrfcuj (mgr.14201) 692 : audit [DBG] from='client.15848 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:13.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:13 vm09 bash[21220]: cluster 2026-03-07T07:07:12.575255+0000 mgr.vm07.yrfcuj (mgr.14201) 693 : cluster [DBG] pgmap v412: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:07:13.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:13 vm09 bash[21220]: cluster 2026-03-07T07:07:12.575255+0000 mgr.vm07.yrfcuj (mgr.14201) 693 : cluster [DBG] pgmap v412: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:07:13.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:13 vm07 bash[17031]: audit 2026-03-07T07:07:12.120464+0000 mgr.vm07.yrfcuj (mgr.14201) 691 : audit [DBG] from='client.15844 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:13.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:13 vm07 bash[17031]: audit 2026-03-07T07:07:12.120464+0000 mgr.vm07.yrfcuj (mgr.14201) 691 : audit [DBG] from='client.15844 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:13.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:13 vm07 bash[17031]: audit 2026-03-07T07:07:12.287734+0000 mgr.vm07.yrfcuj (mgr.14201) 692 : audit [DBG] from='client.15848 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:13.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:13 vm07 bash[17031]: audit 2026-03-07T07:07:12.287734+0000 mgr.vm07.yrfcuj (mgr.14201) 692 : audit [DBG] from='client.15848 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:13.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:13 vm07 bash[17031]: cluster 2026-03-07T07:07:12.575255+0000 mgr.vm07.yrfcuj (mgr.14201) 693 : cluster [DBG] pgmap v412: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:07:13.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:13 vm07 bash[17031]: cluster 2026-03-07T07:07:12.575255+0000 mgr.vm07.yrfcuj (mgr.14201) 693 : cluster [DBG] pgmap v412: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:07:14.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:14 vm07 bash[17031]: cluster 2026-03-07T07:07:14.575778+0000 mgr.vm07.yrfcuj (mgr.14201) 694 : cluster [DBG] pgmap v413: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:14.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:14 vm07 bash[17031]: cluster 2026-03-07T07:07:14.575778+0000 mgr.vm07.yrfcuj (mgr.14201) 694 : cluster [DBG] pgmap v413: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:15.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:14 vm09 bash[21220]: cluster 2026-03-07T07:07:14.575778+0000 mgr.vm07.yrfcuj (mgr.14201) 694 : cluster [DBG] pgmap v413: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:15.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:14 vm09 bash[21220]: cluster 2026-03-07T07:07:14.575778+0000 mgr.vm07.yrfcuj (mgr.14201) 694 : cluster [DBG] pgmap v413: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:16.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:16 vm07 bash[17031]: cluster 2026-03-07T07:07:16.576114+0000 mgr.vm07.yrfcuj (mgr.14201) 695 : cluster [DBG] pgmap v414: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:16.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:16 vm07 bash[17031]: cluster 2026-03-07T07:07:16.576114+0000 mgr.vm07.yrfcuj (mgr.14201) 695 : cluster [DBG] pgmap v414: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:17.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:16 vm09 bash[21220]: cluster 2026-03-07T07:07:16.576114+0000 mgr.vm07.yrfcuj (mgr.14201) 695 : cluster [DBG] pgmap v414: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:17.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:16 vm09 bash[21220]: cluster 2026-03-07T07:07:16.576114+0000 mgr.vm07.yrfcuj (mgr.14201) 695 : cluster [DBG] pgmap v414: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:17.652 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:07:17.802 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:07:17.803 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (3m) 3m ago 9m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:07:17.803 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 3m ago 9m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:07:17.803 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 2m ago 9m - - 2026-03-07T08:07:17.803 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (9m) 2m ago 9m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:07:17.988 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:07:17.988 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:07:17.988 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:07:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:18 vm09 bash[21220]: audit 2026-03-07T07:07:17.990768+0000 mon.vm07 (mon.0) 1061 : audit [DBG] from='client.? 192.168.123.107:0/3539604616' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:18 vm09 bash[21220]: audit 2026-03-07T07:07:17.990768+0000 mon.vm07 (mon.0) 1061 : audit [DBG] from='client.? 192.168.123.107:0/3539604616' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:18.400 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:18 vm07 bash[17031]: audit 2026-03-07T07:07:17.990768+0000 mon.vm07 (mon.0) 1061 : audit [DBG] from='client.? 192.168.123.107:0/3539604616' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:18.400 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:18 vm07 bash[17031]: audit 2026-03-07T07:07:17.990768+0000 mon.vm07 (mon.0) 1061 : audit [DBG] from='client.? 192.168.123.107:0/3539604616' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:19.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:19 vm09 bash[21220]: audit 2026-03-07T07:07:17.639614+0000 mgr.vm07.yrfcuj (mgr.14201) 696 : audit [DBG] from='client.15856 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:19.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:19 vm09 bash[21220]: audit 2026-03-07T07:07:17.639614+0000 mgr.vm07.yrfcuj (mgr.14201) 696 : audit [DBG] from='client.15856 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:19.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:19 vm09 bash[21220]: audit 2026-03-07T07:07:17.803096+0000 mgr.vm07.yrfcuj (mgr.14201) 697 : audit [DBG] from='client.15860 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:19.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:19 vm09 bash[21220]: audit 2026-03-07T07:07:17.803096+0000 mgr.vm07.yrfcuj (mgr.14201) 697 : audit [DBG] from='client.15860 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:19.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:19 vm09 bash[21220]: cluster 2026-03-07T07:07:18.576526+0000 mgr.vm07.yrfcuj (mgr.14201) 698 : cluster [DBG] pgmap v415: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:07:19.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:19 vm09 bash[21220]: cluster 2026-03-07T07:07:18.576526+0000 mgr.vm07.yrfcuj (mgr.14201) 698 : cluster [DBG] pgmap v415: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:07:19.400 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:19 vm07 bash[17031]: audit 2026-03-07T07:07:17.639614+0000 mgr.vm07.yrfcuj (mgr.14201) 696 : audit [DBG] from='client.15856 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:19.400 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:19 vm07 bash[17031]: audit 2026-03-07T07:07:17.639614+0000 mgr.vm07.yrfcuj (mgr.14201) 696 : audit [DBG] from='client.15856 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:19.400 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:19 vm07 bash[17031]: audit 2026-03-07T07:07:17.803096+0000 mgr.vm07.yrfcuj (mgr.14201) 697 : audit [DBG] from='client.15860 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:19.400 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:19 vm07 bash[17031]: audit 2026-03-07T07:07:17.803096+0000 mgr.vm07.yrfcuj (mgr.14201) 697 : audit [DBG] from='client.15860 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:19.400 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:19 vm07 bash[17031]: cluster 2026-03-07T07:07:18.576526+0000 mgr.vm07.yrfcuj (mgr.14201) 698 : cluster [DBG] pgmap v415: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:07:19.400 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:19 vm07 bash[17031]: cluster 2026-03-07T07:07:18.576526+0000 mgr.vm07.yrfcuj (mgr.14201) 698 : cluster [DBG] pgmap v415: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:07:20.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:20 vm07 bash[17031]: cluster 2026-03-07T07:07:20.576880+0000 mgr.vm07.yrfcuj (mgr.14201) 699 : cluster [DBG] pgmap v416: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:07:20.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:20 vm07 bash[17031]: cluster 2026-03-07T07:07:20.576880+0000 mgr.vm07.yrfcuj (mgr.14201) 699 : cluster [DBG] pgmap v416: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:07:21.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:20 vm09 bash[21220]: cluster 2026-03-07T07:07:20.576880+0000 mgr.vm07.yrfcuj (mgr.14201) 699 : cluster [DBG] pgmap v416: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:07:21.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:20 vm09 bash[21220]: cluster 2026-03-07T07:07:20.576880+0000 mgr.vm07.yrfcuj (mgr.14201) 699 : cluster [DBG] pgmap v416: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:07:22.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:22 vm07 bash[17031]: cluster 2026-03-07T07:07:22.577263+0000 mgr.vm07.yrfcuj (mgr.14201) 700 : cluster [DBG] pgmap v417: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:07:22.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:22 vm07 bash[17031]: cluster 2026-03-07T07:07:22.577263+0000 mgr.vm07.yrfcuj (mgr.14201) 700 : cluster [DBG] pgmap v417: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:07:23.092 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:22 vm09 bash[21220]: cluster 2026-03-07T07:07:22.577263+0000 mgr.vm07.yrfcuj (mgr.14201) 700 : cluster [DBG] pgmap v417: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:07:23.092 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:22 vm09 bash[21220]: cluster 2026-03-07T07:07:22.577263+0000 mgr.vm07.yrfcuj (mgr.14201) 700 : cluster [DBG] pgmap v417: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:07:23.163 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:07:23.322 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:07:23.322 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (3m) 3m ago 9m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:07:23.322 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 3m ago 9m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:07:23.322 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 2m ago 9m - - 2026-03-07T08:07:23.322 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (9m) 2m ago 9m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:07:23.521 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:07:23.521 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:07:23.521 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:07:23.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:23 vm07 bash[17031]: audit 2026-03-07T07:07:23.150473+0000 mgr.vm07.yrfcuj (mgr.14201) 701 : audit [DBG] from='client.15868 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:23.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:23 vm07 bash[17031]: audit 2026-03-07T07:07:23.150473+0000 mgr.vm07.yrfcuj (mgr.14201) 701 : audit [DBG] from='client.15868 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:23.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:23 vm07 bash[17031]: audit 2026-03-07T07:07:23.320210+0000 mgr.vm07.yrfcuj (mgr.14201) 702 : audit [DBG] from='client.25215 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:23.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:23 vm07 bash[17031]: audit 2026-03-07T07:07:23.320210+0000 mgr.vm07.yrfcuj (mgr.14201) 702 : audit [DBG] from='client.25215 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:23.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:23 vm07 bash[17031]: audit 2026-03-07T07:07:23.524401+0000 mon.vm07 (mon.0) 1062 : audit [DBG] from='client.? 192.168.123.107:0/1868574800' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:23.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:23 vm07 bash[17031]: audit 2026-03-07T07:07:23.524401+0000 mon.vm07 (mon.0) 1062 : audit [DBG] from='client.? 192.168.123.107:0/1868574800' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:24.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:23 vm09 bash[21220]: audit 2026-03-07T07:07:23.150473+0000 mgr.vm07.yrfcuj (mgr.14201) 701 : audit [DBG] from='client.15868 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:24.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:23 vm09 bash[21220]: audit 2026-03-07T07:07:23.150473+0000 mgr.vm07.yrfcuj (mgr.14201) 701 : audit [DBG] from='client.15868 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:24.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:23 vm09 bash[21220]: audit 2026-03-07T07:07:23.320210+0000 mgr.vm07.yrfcuj (mgr.14201) 702 : audit [DBG] from='client.25215 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:24.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:23 vm09 bash[21220]: audit 2026-03-07T07:07:23.320210+0000 mgr.vm07.yrfcuj (mgr.14201) 702 : audit [DBG] from='client.25215 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:24.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:23 vm09 bash[21220]: audit 2026-03-07T07:07:23.524401+0000 mon.vm07 (mon.0) 1062 : audit [DBG] from='client.? 192.168.123.107:0/1868574800' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:24.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:23 vm09 bash[21220]: audit 2026-03-07T07:07:23.524401+0000 mon.vm07 (mon.0) 1062 : audit [DBG] from='client.? 192.168.123.107:0/1868574800' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:24.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:24 vm07 bash[17031]: cluster 2026-03-07T07:07:24.577755+0000 mgr.vm07.yrfcuj (mgr.14201) 703 : cluster [DBG] pgmap v418: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:24.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:24 vm07 bash[17031]: cluster 2026-03-07T07:07:24.577755+0000 mgr.vm07.yrfcuj (mgr.14201) 703 : cluster [DBG] pgmap v418: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:25.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:24 vm09 bash[21220]: cluster 2026-03-07T07:07:24.577755+0000 mgr.vm07.yrfcuj (mgr.14201) 703 : cluster [DBG] pgmap v418: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:25.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:24 vm09 bash[21220]: cluster 2026-03-07T07:07:24.577755+0000 mgr.vm07.yrfcuj (mgr.14201) 703 : cluster [DBG] pgmap v418: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:25.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:25 vm07 bash[17031]: audit 2026-03-07T07:07:25.138461+0000 mon.vm07 (mon.0) 1063 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:07:25.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:25 vm07 bash[17031]: audit 2026-03-07T07:07:25.138461+0000 mon.vm07 (mon.0) 1063 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:07:26.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:25 vm09 bash[21220]: audit 2026-03-07T07:07:25.138461+0000 mon.vm07 (mon.0) 1063 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:07:26.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:25 vm09 bash[21220]: audit 2026-03-07T07:07:25.138461+0000 mon.vm07 (mon.0) 1063 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:07:26.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:26 vm07 bash[17031]: cluster 2026-03-07T07:07:26.578163+0000 mgr.vm07.yrfcuj (mgr.14201) 704 : cluster [DBG] pgmap v419: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:26.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:26 vm07 bash[17031]: cluster 2026-03-07T07:07:26.578163+0000 mgr.vm07.yrfcuj (mgr.14201) 704 : cluster [DBG] pgmap v419: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:27.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:26 vm09 bash[21220]: cluster 2026-03-07T07:07:26.578163+0000 mgr.vm07.yrfcuj (mgr.14201) 704 : cluster [DBG] pgmap v419: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:27.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:26 vm09 bash[21220]: cluster 2026-03-07T07:07:26.578163+0000 mgr.vm07.yrfcuj (mgr.14201) 704 : cluster [DBG] pgmap v419: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:28.716 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:07:28.870 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:07:28.870 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (4m) 3m ago 9m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:07:28.870 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 3m ago 9m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:07:28.870 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 2m ago 9m - - 2026-03-07T08:07:28.870 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (9m) 2m ago 9m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:07:28.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:28 vm07 bash[17031]: cluster 2026-03-07T07:07:28.578590+0000 mgr.vm07.yrfcuj (mgr.14201) 705 : cluster [DBG] pgmap v420: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:28.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:28 vm07 bash[17031]: cluster 2026-03-07T07:07:28.578590+0000 mgr.vm07.yrfcuj (mgr.14201) 705 : cluster [DBG] pgmap v420: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:29.066 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:07:29.066 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:07:29.066 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:07:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:28 vm09 bash[21220]: cluster 2026-03-07T07:07:28.578590+0000 mgr.vm07.yrfcuj (mgr.14201) 705 : cluster [DBG] pgmap v420: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:28 vm09 bash[21220]: cluster 2026-03-07T07:07:28.578590+0000 mgr.vm07.yrfcuj (mgr.14201) 705 : cluster [DBG] pgmap v420: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:29.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:29 vm07 bash[17031]: audit 2026-03-07T07:07:28.700441+0000 mgr.vm07.yrfcuj (mgr.14201) 706 : audit [DBG] from='client.15880 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:29.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:29 vm07 bash[17031]: audit 2026-03-07T07:07:28.700441+0000 mgr.vm07.yrfcuj (mgr.14201) 706 : audit [DBG] from='client.15880 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:29.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:29 vm07 bash[17031]: audit 2026-03-07T07:07:28.871266+0000 mgr.vm07.yrfcuj (mgr.14201) 707 : audit [DBG] from='client.15884 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:29.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:29 vm07 bash[17031]: audit 2026-03-07T07:07:28.871266+0000 mgr.vm07.yrfcuj (mgr.14201) 707 : audit [DBG] from='client.15884 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:29.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:29 vm07 bash[17031]: audit 2026-03-07T07:07:29.069805+0000 mon.vm07 (mon.0) 1064 : audit [DBG] from='client.? 192.168.123.107:0/690781018' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:29.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:29 vm07 bash[17031]: audit 2026-03-07T07:07:29.069805+0000 mon.vm07 (mon.0) 1064 : audit [DBG] from='client.? 192.168.123.107:0/690781018' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:29 vm09 bash[21220]: audit 2026-03-07T07:07:28.700441+0000 mgr.vm07.yrfcuj (mgr.14201) 706 : audit [DBG] from='client.15880 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:29 vm09 bash[21220]: audit 2026-03-07T07:07:28.700441+0000 mgr.vm07.yrfcuj (mgr.14201) 706 : audit [DBG] from='client.15880 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:29 vm09 bash[21220]: audit 2026-03-07T07:07:28.871266+0000 mgr.vm07.yrfcuj (mgr.14201) 707 : audit [DBG] from='client.15884 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:29 vm09 bash[21220]: audit 2026-03-07T07:07:28.871266+0000 mgr.vm07.yrfcuj (mgr.14201) 707 : audit [DBG] from='client.15884 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:29 vm09 bash[21220]: audit 2026-03-07T07:07:29.069805+0000 mon.vm07 (mon.0) 1064 : audit [DBG] from='client.? 192.168.123.107:0/690781018' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:29 vm09 bash[21220]: audit 2026-03-07T07:07:29.069805+0000 mon.vm07 (mon.0) 1064 : audit [DBG] from='client.? 192.168.123.107:0/690781018' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:30.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:30 vm07 bash[17031]: cluster 2026-03-07T07:07:30.578937+0000 mgr.vm07.yrfcuj (mgr.14201) 708 : cluster [DBG] pgmap v421: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:07:30.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:30 vm07 bash[17031]: cluster 2026-03-07T07:07:30.578937+0000 mgr.vm07.yrfcuj (mgr.14201) 708 : cluster [DBG] pgmap v421: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:07:31.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:30 vm09 bash[21220]: cluster 2026-03-07T07:07:30.578937+0000 mgr.vm07.yrfcuj (mgr.14201) 708 : cluster [DBG] pgmap v421: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:07:31.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:30 vm09 bash[21220]: cluster 2026-03-07T07:07:30.578937+0000 mgr.vm07.yrfcuj (mgr.14201) 708 : cluster [DBG] pgmap v421: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:07:33.092 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:32 vm09 bash[21220]: cluster 2026-03-07T07:07:32.579333+0000 mgr.vm07.yrfcuj (mgr.14201) 709 : cluster [DBG] pgmap v422: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:33.092 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:32 vm09 bash[21220]: cluster 2026-03-07T07:07:32.579333+0000 mgr.vm07.yrfcuj (mgr.14201) 709 : cluster [DBG] pgmap v422: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:33.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:32 vm07 bash[17031]: cluster 2026-03-07T07:07:32.579333+0000 mgr.vm07.yrfcuj (mgr.14201) 709 : cluster [DBG] pgmap v422: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:33.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:32 vm07 bash[17031]: cluster 2026-03-07T07:07:32.579333+0000 mgr.vm07.yrfcuj (mgr.14201) 709 : cluster [DBG] pgmap v422: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:34.238 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:07:34.393 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:07:34.393 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (4m) 3m ago 9m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:07:34.393 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 3m ago 9m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:07:34.393 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 2m ago 9m - - 2026-03-07T08:07:34.393 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (9m) 2m ago 9m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:07:34.585 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:07:34.585 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:07:34.585 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:07:35.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:34 vm09 bash[21220]: audit 2026-03-07T07:07:34.227116+0000 mgr.vm07.yrfcuj (mgr.14201) 710 : audit [DBG] from='client.15892 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:35.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:34 vm09 bash[21220]: audit 2026-03-07T07:07:34.227116+0000 mgr.vm07.yrfcuj (mgr.14201) 710 : audit [DBG] from='client.15892 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:35.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:34 vm09 bash[21220]: audit 2026-03-07T07:07:34.393626+0000 mgr.vm07.yrfcuj (mgr.14201) 711 : audit [DBG] from='client.15896 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:35.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:34 vm09 bash[21220]: audit 2026-03-07T07:07:34.393626+0000 mgr.vm07.yrfcuj (mgr.14201) 711 : audit [DBG] from='client.15896 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:35.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:34 vm09 bash[21220]: cluster 2026-03-07T07:07:34.579761+0000 mgr.vm07.yrfcuj (mgr.14201) 712 : cluster [DBG] pgmap v423: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:35.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:34 vm09 bash[21220]: cluster 2026-03-07T07:07:34.579761+0000 mgr.vm07.yrfcuj (mgr.14201) 712 : cluster [DBG] pgmap v423: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:35.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:34 vm09 bash[21220]: audit 2026-03-07T07:07:34.588093+0000 mon.vm07 (mon.0) 1065 : audit [DBG] from='client.? 192.168.123.107:0/564470541' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:35.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:34 vm09 bash[21220]: audit 2026-03-07T07:07:34.588093+0000 mon.vm07 (mon.0) 1065 : audit [DBG] from='client.? 192.168.123.107:0/564470541' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:35.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:34 vm07 bash[17031]: audit 2026-03-07T07:07:34.227116+0000 mgr.vm07.yrfcuj (mgr.14201) 710 : audit [DBG] from='client.15892 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:35.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:34 vm07 bash[17031]: audit 2026-03-07T07:07:34.227116+0000 mgr.vm07.yrfcuj (mgr.14201) 710 : audit [DBG] from='client.15892 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:35.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:34 vm07 bash[17031]: audit 2026-03-07T07:07:34.393626+0000 mgr.vm07.yrfcuj (mgr.14201) 711 : audit [DBG] from='client.15896 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:35.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:34 vm07 bash[17031]: audit 2026-03-07T07:07:34.393626+0000 mgr.vm07.yrfcuj (mgr.14201) 711 : audit [DBG] from='client.15896 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:35.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:34 vm07 bash[17031]: cluster 2026-03-07T07:07:34.579761+0000 mgr.vm07.yrfcuj (mgr.14201) 712 : cluster [DBG] pgmap v423: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:35.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:34 vm07 bash[17031]: cluster 2026-03-07T07:07:34.579761+0000 mgr.vm07.yrfcuj (mgr.14201) 712 : cluster [DBG] pgmap v423: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:35.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:34 vm07 bash[17031]: audit 2026-03-07T07:07:34.588093+0000 mon.vm07 (mon.0) 1065 : audit [DBG] from='client.? 192.168.123.107:0/564470541' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:35.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:34 vm07 bash[17031]: audit 2026-03-07T07:07:34.588093+0000 mon.vm07 (mon.0) 1065 : audit [DBG] from='client.? 192.168.123.107:0/564470541' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:36.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:36 vm07 bash[17031]: cluster 2026-03-07T07:07:36.580170+0000 mgr.vm07.yrfcuj (mgr.14201) 713 : cluster [DBG] pgmap v424: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:36.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:36 vm07 bash[17031]: cluster 2026-03-07T07:07:36.580170+0000 mgr.vm07.yrfcuj (mgr.14201) 713 : cluster [DBG] pgmap v424: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:36.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:36 vm07 bash[17031]: audit 2026-03-07T07:07:36.614223+0000 mon.vm07 (mon.0) 1066 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:07:36.900 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:36 vm07 bash[17031]: audit 2026-03-07T07:07:36.614223+0000 mon.vm07 (mon.0) 1066 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:07:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:36 vm09 bash[21220]: cluster 2026-03-07T07:07:36.580170+0000 mgr.vm07.yrfcuj (mgr.14201) 713 : cluster [DBG] pgmap v424: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:36 vm09 bash[21220]: cluster 2026-03-07T07:07:36.580170+0000 mgr.vm07.yrfcuj (mgr.14201) 713 : cluster [DBG] pgmap v424: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:07:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:36 vm09 bash[21220]: audit 2026-03-07T07:07:36.614223+0000 mon.vm07 (mon.0) 1066 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:07:37.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:36 vm09 bash[21220]: audit 2026-03-07T07:07:36.614223+0000 mon.vm07 (mon.0) 1066 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:07:38.092 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:37 vm09 bash[21220]: audit 2026-03-07T07:07:36.923418+0000 mon.vm07 (mon.0) 1067 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:07:38.093 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:37 vm09 bash[21220]: audit 2026-03-07T07:07:36.923418+0000 mon.vm07 (mon.0) 1067 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:07:38.093 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:37 vm09 bash[21220]: audit 2026-03-07T07:07:36.923905+0000 mon.vm07 (mon.0) 1068 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:07:38.093 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:37 vm09 bash[21220]: audit 2026-03-07T07:07:36.923905+0000 mon.vm07 (mon.0) 1068 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:07:38.093 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:37 vm09 bash[21220]: cluster 2026-03-07T07:07:36.925045+0000 mgr.vm07.yrfcuj (mgr.14201) 714 : cluster [DBG] pgmap v425: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:38.093 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:37 vm09 bash[21220]: cluster 2026-03-07T07:07:36.925045+0000 mgr.vm07.yrfcuj (mgr.14201) 714 : cluster [DBG] pgmap v425: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:38.093 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:37 vm09 bash[21220]: audit 2026-03-07T07:07:36.928488+0000 mon.vm07 (mon.0) 1069 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:07:38.093 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:37 vm09 bash[21220]: audit 2026-03-07T07:07:36.928488+0000 mon.vm07 (mon.0) 1069 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:07:38.093 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:37 vm09 bash[21220]: audit 2026-03-07T07:07:36.929911+0000 mon.vm07 (mon.0) 1070 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:07:38.093 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:37 vm09 bash[21220]: audit 2026-03-07T07:07:36.929911+0000 mon.vm07 (mon.0) 1070 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:07:38.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:37 vm07 bash[17031]: audit 2026-03-07T07:07:36.923418+0000 mon.vm07 (mon.0) 1067 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:07:38.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:37 vm07 bash[17031]: audit 2026-03-07T07:07:36.923418+0000 mon.vm07 (mon.0) 1067 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:07:38.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:37 vm07 bash[17031]: audit 2026-03-07T07:07:36.923905+0000 mon.vm07 (mon.0) 1068 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:07:38.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:37 vm07 bash[17031]: audit 2026-03-07T07:07:36.923905+0000 mon.vm07 (mon.0) 1068 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:07:38.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:37 vm07 bash[17031]: cluster 2026-03-07T07:07:36.925045+0000 mgr.vm07.yrfcuj (mgr.14201) 714 : cluster [DBG] pgmap v425: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:38.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:37 vm07 bash[17031]: cluster 2026-03-07T07:07:36.925045+0000 mgr.vm07.yrfcuj (mgr.14201) 714 : cluster [DBG] pgmap v425: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:38.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:37 vm07 bash[17031]: audit 2026-03-07T07:07:36.928488+0000 mon.vm07 (mon.0) 1069 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:07:38.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:37 vm07 bash[17031]: audit 2026-03-07T07:07:36.928488+0000 mon.vm07 (mon.0) 1069 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:07:38.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:37 vm07 bash[17031]: audit 2026-03-07T07:07:36.929911+0000 mon.vm07 (mon.0) 1070 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:07:38.150 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:37 vm07 bash[17031]: audit 2026-03-07T07:07:36.929911+0000 mon.vm07 (mon.0) 1070 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:07:39.755 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:07:39.911 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:07:39.911 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (4m) 3m ago 10m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:07:39.911 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 3m ago 10m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:07:39.911 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 3m ago 10m - - 2026-03-07T08:07:39.911 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (10m) 3m ago 10m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:07:40.101 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:07:40.101 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:07:40.101 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:07:40.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:39 vm09 bash[21220]: cluster 2026-03-07T07:07:38.925456+0000 mgr.vm07.yrfcuj (mgr.14201) 715 : cluster [DBG] pgmap v426: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:40.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:39 vm09 bash[21220]: cluster 2026-03-07T07:07:38.925456+0000 mgr.vm07.yrfcuj (mgr.14201) 715 : cluster [DBG] pgmap v426: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:40.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:39 vm07 bash[17031]: cluster 2026-03-07T07:07:38.925456+0000 mgr.vm07.yrfcuj (mgr.14201) 715 : cluster [DBG] pgmap v426: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:40.400 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:39 vm07 bash[17031]: cluster 2026-03-07T07:07:38.925456+0000 mgr.vm07.yrfcuj (mgr.14201) 715 : cluster [DBG] pgmap v426: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:41.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:40 vm09 bash[21220]: audit 2026-03-07T07:07:39.743835+0000 mgr.vm07.yrfcuj (mgr.14201) 716 : audit [DBG] from='client.15904 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:41.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:40 vm09 bash[21220]: audit 2026-03-07T07:07:39.743835+0000 mgr.vm07.yrfcuj (mgr.14201) 716 : audit [DBG] from='client.15904 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:41.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:40 vm09 bash[21220]: audit 2026-03-07T07:07:39.911434+0000 mgr.vm07.yrfcuj (mgr.14201) 717 : audit [DBG] from='client.15908 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:41.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:40 vm09 bash[21220]: audit 2026-03-07T07:07:39.911434+0000 mgr.vm07.yrfcuj (mgr.14201) 717 : audit [DBG] from='client.15908 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:41.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:40 vm09 bash[21220]: audit 2026-03-07T07:07:40.104410+0000 mon.vm07 (mon.0) 1071 : audit [DBG] from='client.? 192.168.123.107:0/3009962630' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:41.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:40 vm09 bash[21220]: audit 2026-03-07T07:07:40.104410+0000 mon.vm07 (mon.0) 1071 : audit [DBG] from='client.? 192.168.123.107:0/3009962630' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:41.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:40 vm09 bash[21220]: audit 2026-03-07T07:07:40.138550+0000 mon.vm07 (mon.0) 1072 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:07:41.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:40 vm09 bash[21220]: audit 2026-03-07T07:07:40.138550+0000 mon.vm07 (mon.0) 1072 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:07:41.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:40 vm07 bash[17031]: audit 2026-03-07T07:07:39.743835+0000 mgr.vm07.yrfcuj (mgr.14201) 716 : audit [DBG] from='client.15904 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:41.400 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:40 vm07 bash[17031]: audit 2026-03-07T07:07:39.743835+0000 mgr.vm07.yrfcuj (mgr.14201) 716 : audit [DBG] from='client.15904 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:41.400 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:40 vm07 bash[17031]: audit 2026-03-07T07:07:39.911434+0000 mgr.vm07.yrfcuj (mgr.14201) 717 : audit [DBG] from='client.15908 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:41.400 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:40 vm07 bash[17031]: audit 2026-03-07T07:07:39.911434+0000 mgr.vm07.yrfcuj (mgr.14201) 717 : audit [DBG] from='client.15908 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:41.400 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:40 vm07 bash[17031]: audit 2026-03-07T07:07:40.104410+0000 mon.vm07 (mon.0) 1071 : audit [DBG] from='client.? 192.168.123.107:0/3009962630' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:41.400 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:40 vm07 bash[17031]: audit 2026-03-07T07:07:40.104410+0000 mon.vm07 (mon.0) 1071 : audit [DBG] from='client.? 192.168.123.107:0/3009962630' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:41.400 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:40 vm07 bash[17031]: audit 2026-03-07T07:07:40.138550+0000 mon.vm07 (mon.0) 1072 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:07:41.400 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:40 vm07 bash[17031]: audit 2026-03-07T07:07:40.138550+0000 mon.vm07 (mon.0) 1072 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:07:42.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:42 vm09 bash[21220]: cluster 2026-03-07T07:07:40.925844+0000 mgr.vm07.yrfcuj (mgr.14201) 718 : cluster [DBG] pgmap v427: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:42.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:42 vm09 bash[21220]: cluster 2026-03-07T07:07:40.925844+0000 mgr.vm07.yrfcuj (mgr.14201) 718 : cluster [DBG] pgmap v427: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:42.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:42 vm07 bash[17031]: cluster 2026-03-07T07:07:40.925844+0000 mgr.vm07.yrfcuj (mgr.14201) 718 : cluster [DBG] pgmap v427: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:42.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:42 vm07 bash[17031]: cluster 2026-03-07T07:07:40.925844+0000 mgr.vm07.yrfcuj (mgr.14201) 718 : cluster [DBG] pgmap v427: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:44 vm09 bash[21220]: cluster 2026-03-07T07:07:42.926408+0000 mgr.vm07.yrfcuj (mgr.14201) 719 : cluster [DBG] pgmap v428: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:44 vm09 bash[21220]: cluster 2026-03-07T07:07:42.926408+0000 mgr.vm07.yrfcuj (mgr.14201) 719 : cluster [DBG] pgmap v428: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:44.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:44 vm07 bash[17031]: cluster 2026-03-07T07:07:42.926408+0000 mgr.vm07.yrfcuj (mgr.14201) 719 : cluster [DBG] pgmap v428: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:44.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:44 vm07 bash[17031]: cluster 2026-03-07T07:07:42.926408+0000 mgr.vm07.yrfcuj (mgr.14201) 719 : cluster [DBG] pgmap v428: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:45.278 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:07:45.430 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:07:45.431 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (4m) 3m ago 10m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:07:45.431 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 3m ago 10m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:07:45.431 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 3m ago 10m - - 2026-03-07T08:07:45.431 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (10m) 3m ago 10m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:07:45.615 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:07:45.615 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:07:45.615 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:07:46.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:46 vm09 bash[21220]: cluster 2026-03-07T07:07:44.926924+0000 mgr.vm07.yrfcuj (mgr.14201) 720 : cluster [DBG] pgmap v429: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:46.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:46 vm09 bash[21220]: cluster 2026-03-07T07:07:44.926924+0000 mgr.vm07.yrfcuj (mgr.14201) 720 : cluster [DBG] pgmap v429: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:46.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:46 vm09 bash[21220]: audit 2026-03-07T07:07:45.266544+0000 mgr.vm07.yrfcuj (mgr.14201) 721 : audit [DBG] from='client.15916 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:46.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:46 vm09 bash[21220]: audit 2026-03-07T07:07:45.266544+0000 mgr.vm07.yrfcuj (mgr.14201) 721 : audit [DBG] from='client.15916 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:46.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:46 vm09 bash[21220]: audit 2026-03-07T07:07:45.431373+0000 mgr.vm07.yrfcuj (mgr.14201) 722 : audit [DBG] from='client.15920 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:46.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:46 vm09 bash[21220]: audit 2026-03-07T07:07:45.431373+0000 mgr.vm07.yrfcuj (mgr.14201) 722 : audit [DBG] from='client.15920 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:46.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:46 vm09 bash[21220]: audit 2026-03-07T07:07:45.618592+0000 mon.vm07 (mon.0) 1073 : audit [DBG] from='client.? 192.168.123.107:0/1922889629' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:46.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:46 vm09 bash[21220]: audit 2026-03-07T07:07:45.618592+0000 mon.vm07 (mon.0) 1073 : audit [DBG] from='client.? 192.168.123.107:0/1922889629' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:46.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:46 vm07 bash[17031]: cluster 2026-03-07T07:07:44.926924+0000 mgr.vm07.yrfcuj (mgr.14201) 720 : cluster [DBG] pgmap v429: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:46.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:46 vm07 bash[17031]: cluster 2026-03-07T07:07:44.926924+0000 mgr.vm07.yrfcuj (mgr.14201) 720 : cluster [DBG] pgmap v429: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:07:46.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:46 vm07 bash[17031]: audit 2026-03-07T07:07:45.266544+0000 mgr.vm07.yrfcuj (mgr.14201) 721 : audit [DBG] from='client.15916 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:46.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:46 vm07 bash[17031]: audit 2026-03-07T07:07:45.266544+0000 mgr.vm07.yrfcuj (mgr.14201) 721 : audit [DBG] from='client.15916 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:46.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:46 vm07 bash[17031]: audit 2026-03-07T07:07:45.431373+0000 mgr.vm07.yrfcuj (mgr.14201) 722 : audit [DBG] from='client.15920 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:46.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:46 vm07 bash[17031]: audit 2026-03-07T07:07:45.431373+0000 mgr.vm07.yrfcuj (mgr.14201) 722 : audit [DBG] from='client.15920 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:46.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:46 vm07 bash[17031]: audit 2026-03-07T07:07:45.618592+0000 mon.vm07 (mon.0) 1073 : audit [DBG] from='client.? 192.168.123.107:0/1922889629' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:46.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:46 vm07 bash[17031]: audit 2026-03-07T07:07:45.618592+0000 mon.vm07 (mon.0) 1073 : audit [DBG] from='client.? 192.168.123.107:0/1922889629' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:48 vm09 bash[21220]: cluster 2026-03-07T07:07:46.927488+0000 mgr.vm07.yrfcuj (mgr.14201) 723 : cluster [DBG] pgmap v430: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 9.8 KiB/s rd, 395 B/s wr, 16 op/s 2026-03-07T08:07:48.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:48 vm09 bash[21220]: cluster 2026-03-07T07:07:46.927488+0000 mgr.vm07.yrfcuj (mgr.14201) 723 : cluster [DBG] pgmap v430: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 9.8 KiB/s rd, 395 B/s wr, 16 op/s 2026-03-07T08:07:48.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:48 vm07 bash[17031]: cluster 2026-03-07T07:07:46.927488+0000 mgr.vm07.yrfcuj (mgr.14201) 723 : cluster [DBG] pgmap v430: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 9.8 KiB/s rd, 395 B/s wr, 16 op/s 2026-03-07T08:07:48.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:48 vm07 bash[17031]: cluster 2026-03-07T07:07:46.927488+0000 mgr.vm07.yrfcuj (mgr.14201) 723 : cluster [DBG] pgmap v430: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 9.8 KiB/s rd, 395 B/s wr, 16 op/s 2026-03-07T08:07:50.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:50 vm09 bash[21220]: cluster 2026-03-07T07:07:48.927890+0000 mgr.vm07.yrfcuj (mgr.14201) 724 : cluster [DBG] pgmap v431: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 26 KiB/s rd, 341 B/s wr, 43 op/s 2026-03-07T08:07:50.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:50 vm09 bash[21220]: cluster 2026-03-07T07:07:48.927890+0000 mgr.vm07.yrfcuj (mgr.14201) 724 : cluster [DBG] pgmap v431: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 26 KiB/s rd, 341 B/s wr, 43 op/s 2026-03-07T08:07:50.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:50 vm07 bash[17031]: cluster 2026-03-07T07:07:48.927890+0000 mgr.vm07.yrfcuj (mgr.14201) 724 : cluster [DBG] pgmap v431: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 26 KiB/s rd, 341 B/s wr, 43 op/s 2026-03-07T08:07:50.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:50 vm07 bash[17031]: cluster 2026-03-07T07:07:48.927890+0000 mgr.vm07.yrfcuj (mgr.14201) 724 : cluster [DBG] pgmap v431: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 26 KiB/s rd, 341 B/s wr, 43 op/s 2026-03-07T08:07:50.790 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:07:50.941 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:07:50.941 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (4m) 3m ago 10m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:07:50.941 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 3m ago 10m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:07:50.941 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 3m ago 10m - - 2026-03-07T08:07:50.941 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (10m) 3m ago 10m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:07:51.128 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:07:51.128 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:07:51.128 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:07:51.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:51 vm09 bash[21220]: audit 2026-03-07T07:07:51.131504+0000 mon.vm07 (mon.0) 1074 : audit [DBG] from='client.? 192.168.123.107:0/436063966' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:51.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:51 vm09 bash[21220]: audit 2026-03-07T07:07:51.131504+0000 mon.vm07 (mon.0) 1074 : audit [DBG] from='client.? 192.168.123.107:0/436063966' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:51.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:51 vm07 bash[17031]: audit 2026-03-07T07:07:51.131504+0000 mon.vm07 (mon.0) 1074 : audit [DBG] from='client.? 192.168.123.107:0/436063966' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:51.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:51 vm07 bash[17031]: audit 2026-03-07T07:07:51.131504+0000 mon.vm07 (mon.0) 1074 : audit [DBG] from='client.? 192.168.123.107:0/436063966' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:52.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:52 vm09 bash[21220]: audit 2026-03-07T07:07:50.778311+0000 mgr.vm07.yrfcuj (mgr.14201) 725 : audit [DBG] from='client.15928 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:52.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:52 vm09 bash[21220]: audit 2026-03-07T07:07:50.778311+0000 mgr.vm07.yrfcuj (mgr.14201) 725 : audit [DBG] from='client.15928 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:52.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:52 vm09 bash[21220]: cluster 2026-03-07T07:07:50.928238+0000 mgr.vm07.yrfcuj (mgr.14201) 726 : cluster [DBG] pgmap v432: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 31 KiB/s rd, 341 B/s wr, 52 op/s 2026-03-07T08:07:52.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:52 vm09 bash[21220]: cluster 2026-03-07T07:07:50.928238+0000 mgr.vm07.yrfcuj (mgr.14201) 726 : cluster [DBG] pgmap v432: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 31 KiB/s rd, 341 B/s wr, 52 op/s 2026-03-07T08:07:52.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:52 vm09 bash[21220]: audit 2026-03-07T07:07:50.941897+0000 mgr.vm07.yrfcuj (mgr.14201) 727 : audit [DBG] from='client.15932 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:52.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:52 vm09 bash[21220]: audit 2026-03-07T07:07:50.941897+0000 mgr.vm07.yrfcuj (mgr.14201) 727 : audit [DBG] from='client.15932 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:52.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:52 vm07 bash[17031]: audit 2026-03-07T07:07:50.778311+0000 mgr.vm07.yrfcuj (mgr.14201) 725 : audit [DBG] from='client.15928 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:52.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:52 vm07 bash[17031]: audit 2026-03-07T07:07:50.778311+0000 mgr.vm07.yrfcuj (mgr.14201) 725 : audit [DBG] from='client.15928 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:52.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:52 vm07 bash[17031]: cluster 2026-03-07T07:07:50.928238+0000 mgr.vm07.yrfcuj (mgr.14201) 726 : cluster [DBG] pgmap v432: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 31 KiB/s rd, 341 B/s wr, 52 op/s 2026-03-07T08:07:52.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:52 vm07 bash[17031]: cluster 2026-03-07T07:07:50.928238+0000 mgr.vm07.yrfcuj (mgr.14201) 726 : cluster [DBG] pgmap v432: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 31 KiB/s rd, 341 B/s wr, 52 op/s 2026-03-07T08:07:52.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:52 vm07 bash[17031]: audit 2026-03-07T07:07:50.941897+0000 mgr.vm07.yrfcuj (mgr.14201) 727 : audit [DBG] from='client.15932 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:52.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:52 vm07 bash[17031]: audit 2026-03-07T07:07:50.941897+0000 mgr.vm07.yrfcuj (mgr.14201) 727 : audit [DBG] from='client.15932 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:54.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:54 vm09 bash[21220]: cluster 2026-03-07T07:07:52.928787+0000 mgr.vm07.yrfcuj (mgr.14201) 728 : cluster [DBG] pgmap v433: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:07:54.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:54 vm09 bash[21220]: cluster 2026-03-07T07:07:52.928787+0000 mgr.vm07.yrfcuj (mgr.14201) 728 : cluster [DBG] pgmap v433: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:07:54.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:54 vm07 bash[17031]: cluster 2026-03-07T07:07:52.928787+0000 mgr.vm07.yrfcuj (mgr.14201) 728 : cluster [DBG] pgmap v433: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:07:54.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:54 vm07 bash[17031]: cluster 2026-03-07T07:07:52.928787+0000 mgr.vm07.yrfcuj (mgr.14201) 728 : cluster [DBG] pgmap v433: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:07:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:55 vm09 bash[21220]: audit 2026-03-07T07:07:55.138700+0000 mon.vm07 (mon.0) 1075 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:07:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:55 vm09 bash[21220]: audit 2026-03-07T07:07:55.138700+0000 mon.vm07 (mon.0) 1075 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:07:55.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:55 vm07 bash[17031]: audit 2026-03-07T07:07:55.138700+0000 mon.vm07 (mon.0) 1075 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:07:55.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:55 vm07 bash[17031]: audit 2026-03-07T07:07:55.138700+0000 mon.vm07 (mon.0) 1075 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:07:56.302 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:07:56.453 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:07:56.453 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (4m) 3m ago 10m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:07:56.453 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 3m ago 10m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:07:56.453 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 3m ago 10m - - 2026-03-07T08:07:56.453 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (10m) 3m ago 10m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:07:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:56 vm09 bash[21220]: cluster 2026-03-07T07:07:54.929203+0000 mgr.vm07.yrfcuj (mgr.14201) 729 : cluster [DBG] pgmap v434: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:07:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:56 vm09 bash[21220]: cluster 2026-03-07T07:07:54.929203+0000 mgr.vm07.yrfcuj (mgr.14201) 729 : cluster [DBG] pgmap v434: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:07:56.631 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:07:56.631 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:07:56.631 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:07:56.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:56 vm07 bash[17031]: cluster 2026-03-07T07:07:54.929203+0000 mgr.vm07.yrfcuj (mgr.14201) 729 : cluster [DBG] pgmap v434: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:07:56.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:56 vm07 bash[17031]: cluster 2026-03-07T07:07:54.929203+0000 mgr.vm07.yrfcuj (mgr.14201) 729 : cluster [DBG] pgmap v434: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:07:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:57 vm09 bash[21220]: audit 2026-03-07T07:07:56.289971+0000 mgr.vm07.yrfcuj (mgr.14201) 730 : audit [DBG] from='client.15940 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:57 vm09 bash[21220]: audit 2026-03-07T07:07:56.289971+0000 mgr.vm07.yrfcuj (mgr.14201) 730 : audit [DBG] from='client.15940 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:57 vm09 bash[21220]: audit 2026-03-07T07:07:56.453737+0000 mgr.vm07.yrfcuj (mgr.14201) 731 : audit [DBG] from='client.15944 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:57 vm09 bash[21220]: audit 2026-03-07T07:07:56.453737+0000 mgr.vm07.yrfcuj (mgr.14201) 731 : audit [DBG] from='client.15944 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:57 vm09 bash[21220]: audit 2026-03-07T07:07:56.635070+0000 mon.vm07 (mon.0) 1076 : audit [DBG] from='client.? 192.168.123.107:0/793885902' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:57 vm09 bash[21220]: audit 2026-03-07T07:07:56.635070+0000 mon.vm07 (mon.0) 1076 : audit [DBG] from='client.? 192.168.123.107:0/793885902' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:57.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:57 vm07 bash[17031]: audit 2026-03-07T07:07:56.289971+0000 mgr.vm07.yrfcuj (mgr.14201) 730 : audit [DBG] from='client.15940 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:57.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:57 vm07 bash[17031]: audit 2026-03-07T07:07:56.289971+0000 mgr.vm07.yrfcuj (mgr.14201) 730 : audit [DBG] from='client.15940 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:57.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:57 vm07 bash[17031]: audit 2026-03-07T07:07:56.453737+0000 mgr.vm07.yrfcuj (mgr.14201) 731 : audit [DBG] from='client.15944 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:57.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:57 vm07 bash[17031]: audit 2026-03-07T07:07:56.453737+0000 mgr.vm07.yrfcuj (mgr.14201) 731 : audit [DBG] from='client.15944 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:07:57.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:57 vm07 bash[17031]: audit 2026-03-07T07:07:56.635070+0000 mon.vm07 (mon.0) 1076 : audit [DBG] from='client.? 192.168.123.107:0/793885902' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:57.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:57 vm07 bash[17031]: audit 2026-03-07T07:07:56.635070+0000 mon.vm07 (mon.0) 1076 : audit [DBG] from='client.? 192.168.123.107:0/793885902' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:07:58.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:58 vm09 bash[21220]: cluster 2026-03-07T07:07:56.929626+0000 mgr.vm07.yrfcuj (mgr.14201) 732 : cluster [DBG] pgmap v435: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:07:58.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:07:58 vm09 bash[21220]: cluster 2026-03-07T07:07:56.929626+0000 mgr.vm07.yrfcuj (mgr.14201) 732 : cluster [DBG] pgmap v435: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:07:58.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:58 vm07 bash[17031]: cluster 2026-03-07T07:07:56.929626+0000 mgr.vm07.yrfcuj (mgr.14201) 732 : cluster [DBG] pgmap v435: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:07:58.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:07:58 vm07 bash[17031]: cluster 2026-03-07T07:07:56.929626+0000 mgr.vm07.yrfcuj (mgr.14201) 732 : cluster [DBG] pgmap v435: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:08:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:00 vm09 bash[21220]: cluster 2026-03-07T07:07:58.930048+0000 mgr.vm07.yrfcuj (mgr.14201) 733 : cluster [DBG] pgmap v436: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s rd, 0 B/s wr, 45 op/s 2026-03-07T08:08:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:00 vm09 bash[21220]: cluster 2026-03-07T07:07:58.930048+0000 mgr.vm07.yrfcuj (mgr.14201) 733 : cluster [DBG] pgmap v436: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s rd, 0 B/s wr, 45 op/s 2026-03-07T08:08:00.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:00 vm07 bash[17031]: cluster 2026-03-07T07:07:58.930048+0000 mgr.vm07.yrfcuj (mgr.14201) 733 : cluster [DBG] pgmap v436: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s rd, 0 B/s wr, 45 op/s 2026-03-07T08:08:00.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:00 vm07 bash[17031]: cluster 2026-03-07T07:07:58.930048+0000 mgr.vm07.yrfcuj (mgr.14201) 733 : cluster [DBG] pgmap v436: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s rd, 0 B/s wr, 45 op/s 2026-03-07T08:08:01.797 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:08:01.942 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:08:01.942 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (4m) 3m ago 10m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:08:01.942 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (3m) 3m ago 10m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:08:01.942 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 3m ago 10m - - 2026-03-07T08:08:01.942 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (10m) 3m ago 10m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:08:02.126 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:08:02.127 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:08:02.127 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:08:02.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:02 vm09 bash[21220]: cluster 2026-03-07T07:08:00.930434+0000 mgr.vm07.yrfcuj (mgr.14201) 734 : cluster [DBG] pgmap v437: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 10 KiB/s rd, 0 B/s wr, 16 op/s 2026-03-07T08:08:02.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:02 vm09 bash[21220]: cluster 2026-03-07T07:08:00.930434+0000 mgr.vm07.yrfcuj (mgr.14201) 734 : cluster [DBG] pgmap v437: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 10 KiB/s rd, 0 B/s wr, 16 op/s 2026-03-07T08:08:02.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:02 vm09 bash[21220]: audit 2026-03-07T07:08:02.130279+0000 mon.vm07 (mon.0) 1077 : audit [DBG] from='client.? 192.168.123.107:0/3948467417' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:02.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:02 vm09 bash[21220]: audit 2026-03-07T07:08:02.130279+0000 mon.vm07 (mon.0) 1077 : audit [DBG] from='client.? 192.168.123.107:0/3948467417' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:02.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:02 vm07 bash[17031]: cluster 2026-03-07T07:08:00.930434+0000 mgr.vm07.yrfcuj (mgr.14201) 734 : cluster [DBG] pgmap v437: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 10 KiB/s rd, 0 B/s wr, 16 op/s 2026-03-07T08:08:02.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:02 vm07 bash[17031]: cluster 2026-03-07T07:08:00.930434+0000 mgr.vm07.yrfcuj (mgr.14201) 734 : cluster [DBG] pgmap v437: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 10 KiB/s rd, 0 B/s wr, 16 op/s 2026-03-07T08:08:02.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:02 vm07 bash[17031]: audit 2026-03-07T07:08:02.130279+0000 mon.vm07 (mon.0) 1077 : audit [DBG] from='client.? 192.168.123.107:0/3948467417' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:02.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:02 vm07 bash[17031]: audit 2026-03-07T07:08:02.130279+0000 mon.vm07 (mon.0) 1077 : audit [DBG] from='client.? 192.168.123.107:0/3948467417' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:03 vm09 bash[21220]: audit 2026-03-07T07:08:01.786166+0000 mgr.vm07.yrfcuj (mgr.14201) 735 : audit [DBG] from='client.15952 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:03 vm09 bash[21220]: audit 2026-03-07T07:08:01.786166+0000 mgr.vm07.yrfcuj (mgr.14201) 735 : audit [DBG] from='client.15952 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:03 vm09 bash[21220]: audit 2026-03-07T07:08:01.943025+0000 mgr.vm07.yrfcuj (mgr.14201) 736 : audit [DBG] from='client.15956 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:03 vm09 bash[21220]: audit 2026-03-07T07:08:01.943025+0000 mgr.vm07.yrfcuj (mgr.14201) 736 : audit [DBG] from='client.15956 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:03.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:03 vm07 bash[17031]: audit 2026-03-07T07:08:01.786166+0000 mgr.vm07.yrfcuj (mgr.14201) 735 : audit [DBG] from='client.15952 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:03.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:03 vm07 bash[17031]: audit 2026-03-07T07:08:01.786166+0000 mgr.vm07.yrfcuj (mgr.14201) 735 : audit [DBG] from='client.15952 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:03.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:03 vm07 bash[17031]: audit 2026-03-07T07:08:01.943025+0000 mgr.vm07.yrfcuj (mgr.14201) 736 : audit [DBG] from='client.15956 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:03.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:03 vm07 bash[17031]: audit 2026-03-07T07:08:01.943025+0000 mgr.vm07.yrfcuj (mgr.14201) 736 : audit [DBG] from='client.15956 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:04.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:04 vm07 bash[17031]: cluster 2026-03-07T07:08:02.930912+0000 mgr.vm07.yrfcuj (mgr.14201) 737 : cluster [DBG] pgmap v438: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 4.8 KiB/s rd, 170 B/s wr, 8 op/s 2026-03-07T08:08:04.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:04 vm07 bash[17031]: cluster 2026-03-07T07:08:02.930912+0000 mgr.vm07.yrfcuj (mgr.14201) 737 : cluster [DBG] pgmap v438: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 4.8 KiB/s rd, 170 B/s wr, 8 op/s 2026-03-07T08:08:04.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:04 vm09 bash[21220]: cluster 2026-03-07T07:08:02.930912+0000 mgr.vm07.yrfcuj (mgr.14201) 737 : cluster [DBG] pgmap v438: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 4.8 KiB/s rd, 170 B/s wr, 8 op/s 2026-03-07T08:08:04.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:04 vm09 bash[21220]: cluster 2026-03-07T07:08:02.930912+0000 mgr.vm07.yrfcuj (mgr.14201) 737 : cluster [DBG] pgmap v438: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 4.8 KiB/s rd, 170 B/s wr, 8 op/s 2026-03-07T08:08:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:06 vm09 bash[21220]: cluster 2026-03-07T07:08:04.931250+0000 mgr.vm07.yrfcuj (mgr.14201) 738 : cluster [DBG] pgmap v439: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:06 vm09 bash[21220]: cluster 2026-03-07T07:08:04.931250+0000 mgr.vm07.yrfcuj (mgr.14201) 738 : cluster [DBG] pgmap v439: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:06.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:06 vm07 bash[17031]: cluster 2026-03-07T07:08:04.931250+0000 mgr.vm07.yrfcuj (mgr.14201) 738 : cluster [DBG] pgmap v439: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:06.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:06 vm07 bash[17031]: cluster 2026-03-07T07:08:04.931250+0000 mgr.vm07.yrfcuj (mgr.14201) 738 : cluster [DBG] pgmap v439: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:07.307 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:08:07.456 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:08:07.457 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (4m) 3m ago 10m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:08:07.457 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 3m ago 10m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:08:07.457 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 3m ago 10m - - 2026-03-07T08:08:07.457 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (10m) 3m ago 10m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:08:07.645 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:08:07.646 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:08:07.646 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:08:08.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:08 vm09 bash[21220]: cluster 2026-03-07T07:08:06.931615+0000 mgr.vm07.yrfcuj (mgr.14201) 739 : cluster [DBG] pgmap v440: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:08.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:08 vm09 bash[21220]: cluster 2026-03-07T07:08:06.931615+0000 mgr.vm07.yrfcuj (mgr.14201) 739 : cluster [DBG] pgmap v440: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:08.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:08 vm09 bash[21220]: audit 2026-03-07T07:08:07.294766+0000 mgr.vm07.yrfcuj (mgr.14201) 740 : audit [DBG] from='client.15964 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:08.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:08 vm09 bash[21220]: audit 2026-03-07T07:08:07.294766+0000 mgr.vm07.yrfcuj (mgr.14201) 740 : audit [DBG] from='client.15964 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:08.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:08 vm09 bash[21220]: audit 2026-03-07T07:08:07.457882+0000 mgr.vm07.yrfcuj (mgr.14201) 741 : audit [DBG] from='client.15968 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:08.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:08 vm09 bash[21220]: audit 2026-03-07T07:08:07.457882+0000 mgr.vm07.yrfcuj (mgr.14201) 741 : audit [DBG] from='client.15968 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:08.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:08 vm09 bash[21220]: audit 2026-03-07T07:08:07.649021+0000 mon.vm07 (mon.0) 1078 : audit [DBG] from='client.? 192.168.123.107:0/645474313' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:08.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:08 vm09 bash[21220]: audit 2026-03-07T07:08:07.649021+0000 mon.vm07 (mon.0) 1078 : audit [DBG] from='client.? 192.168.123.107:0/645474313' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:08.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:08 vm07 bash[17031]: cluster 2026-03-07T07:08:06.931615+0000 mgr.vm07.yrfcuj (mgr.14201) 739 : cluster [DBG] pgmap v440: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:08.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:08 vm07 bash[17031]: cluster 2026-03-07T07:08:06.931615+0000 mgr.vm07.yrfcuj (mgr.14201) 739 : cluster [DBG] pgmap v440: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:08.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:08 vm07 bash[17031]: audit 2026-03-07T07:08:07.294766+0000 mgr.vm07.yrfcuj (mgr.14201) 740 : audit [DBG] from='client.15964 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:08.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:08 vm07 bash[17031]: audit 2026-03-07T07:08:07.294766+0000 mgr.vm07.yrfcuj (mgr.14201) 740 : audit [DBG] from='client.15964 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:08.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:08 vm07 bash[17031]: audit 2026-03-07T07:08:07.457882+0000 mgr.vm07.yrfcuj (mgr.14201) 741 : audit [DBG] from='client.15968 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:08.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:08 vm07 bash[17031]: audit 2026-03-07T07:08:07.457882+0000 mgr.vm07.yrfcuj (mgr.14201) 741 : audit [DBG] from='client.15968 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:08.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:08 vm07 bash[17031]: audit 2026-03-07T07:08:07.649021+0000 mon.vm07 (mon.0) 1078 : audit [DBG] from='client.? 192.168.123.107:0/645474313' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:08.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:08 vm07 bash[17031]: audit 2026-03-07T07:08:07.649021+0000 mon.vm07 (mon.0) 1078 : audit [DBG] from='client.? 192.168.123.107:0/645474313' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:10 vm09 bash[21220]: cluster 2026-03-07T07:08:08.932048+0000 mgr.vm07.yrfcuj (mgr.14201) 742 : cluster [DBG] pgmap v441: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:10 vm09 bash[21220]: cluster 2026-03-07T07:08:08.932048+0000 mgr.vm07.yrfcuj (mgr.14201) 742 : cluster [DBG] pgmap v441: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:10 vm09 bash[21220]: audit 2026-03-07T07:08:10.138790+0000 mon.vm07 (mon.0) 1079 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:08:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:10 vm09 bash[21220]: audit 2026-03-07T07:08:10.138790+0000 mon.vm07 (mon.0) 1079 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:08:10.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:10 vm07 bash[17031]: cluster 2026-03-07T07:08:08.932048+0000 mgr.vm07.yrfcuj (mgr.14201) 742 : cluster [DBG] pgmap v441: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:10.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:10 vm07 bash[17031]: cluster 2026-03-07T07:08:08.932048+0000 mgr.vm07.yrfcuj (mgr.14201) 742 : cluster [DBG] pgmap v441: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:10.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:10 vm07 bash[17031]: audit 2026-03-07T07:08:10.138790+0000 mon.vm07 (mon.0) 1079 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:08:10.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:10 vm07 bash[17031]: audit 2026-03-07T07:08:10.138790+0000 mon.vm07 (mon.0) 1079 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:08:12.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:12 vm09 bash[21220]: cluster 2026-03-07T07:08:10.932446+0000 mgr.vm07.yrfcuj (mgr.14201) 743 : cluster [DBG] pgmap v442: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:12.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:12 vm09 bash[21220]: cluster 2026-03-07T07:08:10.932446+0000 mgr.vm07.yrfcuj (mgr.14201) 743 : cluster [DBG] pgmap v442: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:12.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:12 vm07 bash[17031]: cluster 2026-03-07T07:08:10.932446+0000 mgr.vm07.yrfcuj (mgr.14201) 743 : cluster [DBG] pgmap v442: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:12.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:12 vm07 bash[17031]: cluster 2026-03-07T07:08:10.932446+0000 mgr.vm07.yrfcuj (mgr.14201) 743 : cluster [DBG] pgmap v442: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:12.817 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:08:12.969 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:08:12.970 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (4m) 4m ago 10m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:08:12.970 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 4m ago 10m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:08:12.970 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 3m ago 10m - - 2026-03-07T08:08:12.970 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (10m) 3m ago 10m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:08:13.154 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:08:13.154 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:08:13.154 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:08:13.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:13 vm09 bash[21220]: audit 2026-03-07T07:08:13.152751+0000 mon.vm09 (mon.1) 45 : audit [DBG] from='client.? 192.168.123.107:0/2284743613' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:13.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:13 vm09 bash[21220]: audit 2026-03-07T07:08:13.152751+0000 mon.vm09 (mon.1) 45 : audit [DBG] from='client.? 192.168.123.107:0/2284743613' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:13.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:13 vm07 bash[17031]: audit 2026-03-07T07:08:13.152751+0000 mon.vm09 (mon.1) 45 : audit [DBG] from='client.? 192.168.123.107:0/2284743613' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:13.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:13 vm07 bash[17031]: audit 2026-03-07T07:08:13.152751+0000 mon.vm09 (mon.1) 45 : audit [DBG] from='client.? 192.168.123.107:0/2284743613' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:14.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:14 vm09 bash[21220]: audit 2026-03-07T07:08:12.805222+0000 mgr.vm07.yrfcuj (mgr.14201) 744 : audit [DBG] from='client.15976 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:14.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:14 vm09 bash[21220]: audit 2026-03-07T07:08:12.805222+0000 mgr.vm07.yrfcuj (mgr.14201) 744 : audit [DBG] from='client.15976 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:14.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:14 vm09 bash[21220]: cluster 2026-03-07T07:08:12.932927+0000 mgr.vm07.yrfcuj (mgr.14201) 745 : cluster [DBG] pgmap v443: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:14.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:14 vm09 bash[21220]: cluster 2026-03-07T07:08:12.932927+0000 mgr.vm07.yrfcuj (mgr.14201) 745 : cluster [DBG] pgmap v443: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:14.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:14 vm09 bash[21220]: audit 2026-03-07T07:08:12.970549+0000 mgr.vm07.yrfcuj (mgr.14201) 746 : audit [DBG] from='client.15980 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:14.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:14 vm09 bash[21220]: audit 2026-03-07T07:08:12.970549+0000 mgr.vm07.yrfcuj (mgr.14201) 746 : audit [DBG] from='client.15980 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:14.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:14 vm07 bash[17031]: audit 2026-03-07T07:08:12.805222+0000 mgr.vm07.yrfcuj (mgr.14201) 744 : audit [DBG] from='client.15976 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:14.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:14 vm07 bash[17031]: audit 2026-03-07T07:08:12.805222+0000 mgr.vm07.yrfcuj (mgr.14201) 744 : audit [DBG] from='client.15976 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:14.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:14 vm07 bash[17031]: cluster 2026-03-07T07:08:12.932927+0000 mgr.vm07.yrfcuj (mgr.14201) 745 : cluster [DBG] pgmap v443: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:14.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:14 vm07 bash[17031]: cluster 2026-03-07T07:08:12.932927+0000 mgr.vm07.yrfcuj (mgr.14201) 745 : cluster [DBG] pgmap v443: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:14.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:14 vm07 bash[17031]: audit 2026-03-07T07:08:12.970549+0000 mgr.vm07.yrfcuj (mgr.14201) 746 : audit [DBG] from='client.15980 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:14.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:14 vm07 bash[17031]: audit 2026-03-07T07:08:12.970549+0000 mgr.vm07.yrfcuj (mgr.14201) 746 : audit [DBG] from='client.15980 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:16.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:16 vm09 bash[21220]: cluster 2026-03-07T07:08:14.933345+0000 mgr.vm07.yrfcuj (mgr.14201) 747 : cluster [DBG] pgmap v444: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:16.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:16 vm09 bash[21220]: cluster 2026-03-07T07:08:14.933345+0000 mgr.vm07.yrfcuj (mgr.14201) 747 : cluster [DBG] pgmap v444: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:16 vm07 bash[17031]: cluster 2026-03-07T07:08:14.933345+0000 mgr.vm07.yrfcuj (mgr.14201) 747 : cluster [DBG] pgmap v444: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:16 vm07 bash[17031]: cluster 2026-03-07T07:08:14.933345+0000 mgr.vm07.yrfcuj (mgr.14201) 747 : cluster [DBG] pgmap v444: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:18.334 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:08:18.482 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:08:18.482 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (4m) 4m ago 10m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:08:18.482 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 4m ago 10m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:08:18.482 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 3m ago 10m - - 2026-03-07T08:08:18.482 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (10m) 3m ago 10m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:08:18.672 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:08:18.672 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:08:18.672 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:08:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:18 vm09 bash[21220]: cluster 2026-03-07T07:08:16.933853+0000 mgr.vm07.yrfcuj (mgr.14201) 748 : cluster [DBG] pgmap v445: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:18 vm09 bash[21220]: cluster 2026-03-07T07:08:16.933853+0000 mgr.vm07.yrfcuj (mgr.14201) 748 : cluster [DBG] pgmap v445: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:18.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:18 vm07 bash[17031]: cluster 2026-03-07T07:08:16.933853+0000 mgr.vm07.yrfcuj (mgr.14201) 748 : cluster [DBG] pgmap v445: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:18.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:18 vm07 bash[17031]: cluster 2026-03-07T07:08:16.933853+0000 mgr.vm07.yrfcuj (mgr.14201) 748 : cluster [DBG] pgmap v445: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:19 vm09 bash[21220]: audit 2026-03-07T07:08:18.322600+0000 mgr.vm07.yrfcuj (mgr.14201) 749 : audit [DBG] from='client.15988 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:19 vm09 bash[21220]: audit 2026-03-07T07:08:18.322600+0000 mgr.vm07.yrfcuj (mgr.14201) 749 : audit [DBG] from='client.15988 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:19 vm09 bash[21220]: audit 2026-03-07T07:08:18.483027+0000 mgr.vm07.yrfcuj (mgr.14201) 750 : audit [DBG] from='client.25305 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:19 vm09 bash[21220]: audit 2026-03-07T07:08:18.483027+0000 mgr.vm07.yrfcuj (mgr.14201) 750 : audit [DBG] from='client.25305 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:19 vm09 bash[21220]: audit 2026-03-07T07:08:18.675272+0000 mon.vm07 (mon.0) 1080 : audit [DBG] from='client.? 192.168.123.107:0/586688031' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:19 vm09 bash[21220]: audit 2026-03-07T07:08:18.675272+0000 mon.vm07 (mon.0) 1080 : audit [DBG] from='client.? 192.168.123.107:0/586688031' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:19.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:19 vm07 bash[17031]: audit 2026-03-07T07:08:18.322600+0000 mgr.vm07.yrfcuj (mgr.14201) 749 : audit [DBG] from='client.15988 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:19.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:19 vm07 bash[17031]: audit 2026-03-07T07:08:18.322600+0000 mgr.vm07.yrfcuj (mgr.14201) 749 : audit [DBG] from='client.15988 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:19.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:19 vm07 bash[17031]: audit 2026-03-07T07:08:18.483027+0000 mgr.vm07.yrfcuj (mgr.14201) 750 : audit [DBG] from='client.25305 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:19.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:19 vm07 bash[17031]: audit 2026-03-07T07:08:18.483027+0000 mgr.vm07.yrfcuj (mgr.14201) 750 : audit [DBG] from='client.25305 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:19.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:19 vm07 bash[17031]: audit 2026-03-07T07:08:18.675272+0000 mon.vm07 (mon.0) 1080 : audit [DBG] from='client.? 192.168.123.107:0/586688031' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:19.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:19 vm07 bash[17031]: audit 2026-03-07T07:08:18.675272+0000 mon.vm07 (mon.0) 1080 : audit [DBG] from='client.? 192.168.123.107:0/586688031' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:20 vm09 bash[21220]: cluster 2026-03-07T07:08:18.934375+0000 mgr.vm07.yrfcuj (mgr.14201) 751 : cluster [DBG] pgmap v446: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:20 vm09 bash[21220]: cluster 2026-03-07T07:08:18.934375+0000 mgr.vm07.yrfcuj (mgr.14201) 751 : cluster [DBG] pgmap v446: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:20.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:20 vm07 bash[17031]: cluster 2026-03-07T07:08:18.934375+0000 mgr.vm07.yrfcuj (mgr.14201) 751 : cluster [DBG] pgmap v446: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:20.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:20 vm07 bash[17031]: cluster 2026-03-07T07:08:18.934375+0000 mgr.vm07.yrfcuj (mgr.14201) 751 : cluster [DBG] pgmap v446: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:22.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:22 vm09 bash[21220]: cluster 2026-03-07T07:08:20.934759+0000 mgr.vm07.yrfcuj (mgr.14201) 752 : cluster [DBG] pgmap v447: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:22.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:22 vm09 bash[21220]: cluster 2026-03-07T07:08:20.934759+0000 mgr.vm07.yrfcuj (mgr.14201) 752 : cluster [DBG] pgmap v447: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:22.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:22 vm07 bash[17031]: cluster 2026-03-07T07:08:20.934759+0000 mgr.vm07.yrfcuj (mgr.14201) 752 : cluster [DBG] pgmap v447: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:22.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:22 vm07 bash[17031]: cluster 2026-03-07T07:08:20.934759+0000 mgr.vm07.yrfcuj (mgr.14201) 752 : cluster [DBG] pgmap v447: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:23.847 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:08:23.993 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:08:23.993 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (4m) 4m ago 10m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:08:23.993 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 4m ago 10m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:08:23.993 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 3m ago 10m - - 2026-03-07T08:08:23.993 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (10m) 3m ago 10m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:08:24.172 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:08:24.172 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:08:24.172 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:08:24.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:24 vm07 bash[17031]: cluster 2026-03-07T07:08:22.935140+0000 mgr.vm07.yrfcuj (mgr.14201) 753 : cluster [DBG] pgmap v448: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:24.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:24 vm07 bash[17031]: cluster 2026-03-07T07:08:22.935140+0000 mgr.vm07.yrfcuj (mgr.14201) 753 : cluster [DBG] pgmap v448: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:24.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:24 vm07 bash[17031]: audit 2026-03-07T07:08:24.171320+0000 mon.vm09 (mon.1) 46 : audit [DBG] from='client.? 192.168.123.107:0/2116735477' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:24.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:24 vm07 bash[17031]: audit 2026-03-07T07:08:24.171320+0000 mon.vm09 (mon.1) 46 : audit [DBG] from='client.? 192.168.123.107:0/2116735477' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:24.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:24 vm09 bash[21220]: cluster 2026-03-07T07:08:22.935140+0000 mgr.vm07.yrfcuj (mgr.14201) 753 : cluster [DBG] pgmap v448: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:24.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:24 vm09 bash[21220]: cluster 2026-03-07T07:08:22.935140+0000 mgr.vm07.yrfcuj (mgr.14201) 753 : cluster [DBG] pgmap v448: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:24.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:24 vm09 bash[21220]: audit 2026-03-07T07:08:24.171320+0000 mon.vm09 (mon.1) 46 : audit [DBG] from='client.? 192.168.123.107:0/2116735477' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:24.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:24 vm09 bash[21220]: audit 2026-03-07T07:08:24.171320+0000 mon.vm09 (mon.1) 46 : audit [DBG] from='client.? 192.168.123.107:0/2116735477' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:25 vm09 bash[21220]: audit 2026-03-07T07:08:23.836452+0000 mgr.vm07.yrfcuj (mgr.14201) 754 : audit [DBG] from='client.16000 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:25 vm09 bash[21220]: audit 2026-03-07T07:08:23.836452+0000 mgr.vm07.yrfcuj (mgr.14201) 754 : audit [DBG] from='client.16000 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:25 vm09 bash[21220]: audit 2026-03-07T07:08:23.993957+0000 mgr.vm07.yrfcuj (mgr.14201) 755 : audit [DBG] from='client.16004 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:25 vm09 bash[21220]: audit 2026-03-07T07:08:23.993957+0000 mgr.vm07.yrfcuj (mgr.14201) 755 : audit [DBG] from='client.16004 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:25 vm09 bash[21220]: audit 2026-03-07T07:08:25.139491+0000 mon.vm07 (mon.0) 1081 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:08:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:25 vm09 bash[21220]: audit 2026-03-07T07:08:25.139491+0000 mon.vm07 (mon.0) 1081 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:08:25.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:25 vm07 bash[17031]: audit 2026-03-07T07:08:23.836452+0000 mgr.vm07.yrfcuj (mgr.14201) 754 : audit [DBG] from='client.16000 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:25.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:25 vm07 bash[17031]: audit 2026-03-07T07:08:23.836452+0000 mgr.vm07.yrfcuj (mgr.14201) 754 : audit [DBG] from='client.16000 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:25.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:25 vm07 bash[17031]: audit 2026-03-07T07:08:23.993957+0000 mgr.vm07.yrfcuj (mgr.14201) 755 : audit [DBG] from='client.16004 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:25.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:25 vm07 bash[17031]: audit 2026-03-07T07:08:23.993957+0000 mgr.vm07.yrfcuj (mgr.14201) 755 : audit [DBG] from='client.16004 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:25.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:25 vm07 bash[17031]: audit 2026-03-07T07:08:25.139491+0000 mon.vm07 (mon.0) 1081 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:08:25.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:25 vm07 bash[17031]: audit 2026-03-07T07:08:25.139491+0000 mon.vm07 (mon.0) 1081 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:08:26.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:26 vm09 bash[21220]: cluster 2026-03-07T07:08:24.935508+0000 mgr.vm07.yrfcuj (mgr.14201) 756 : cluster [DBG] pgmap v449: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:26.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:26 vm09 bash[21220]: cluster 2026-03-07T07:08:24.935508+0000 mgr.vm07.yrfcuj (mgr.14201) 756 : cluster [DBG] pgmap v449: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:26.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:26 vm07 bash[17031]: cluster 2026-03-07T07:08:24.935508+0000 mgr.vm07.yrfcuj (mgr.14201) 756 : cluster [DBG] pgmap v449: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:26.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:26 vm07 bash[17031]: cluster 2026-03-07T07:08:24.935508+0000 mgr.vm07.yrfcuj (mgr.14201) 756 : cluster [DBG] pgmap v449: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:28 vm09 bash[21220]: cluster 2026-03-07T07:08:26.935868+0000 mgr.vm07.yrfcuj (mgr.14201) 757 : cluster [DBG] pgmap v450: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:28 vm09 bash[21220]: cluster 2026-03-07T07:08:26.935868+0000 mgr.vm07.yrfcuj (mgr.14201) 757 : cluster [DBG] pgmap v450: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:28.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:28 vm07 bash[17031]: cluster 2026-03-07T07:08:26.935868+0000 mgr.vm07.yrfcuj (mgr.14201) 757 : cluster [DBG] pgmap v450: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:28.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:28 vm07 bash[17031]: cluster 2026-03-07T07:08:26.935868+0000 mgr.vm07.yrfcuj (mgr.14201) 757 : cluster [DBG] pgmap v450: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:29.346 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:08:29.512 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:08:29.512 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (5m) 4m ago 10m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:08:29.512 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 4m ago 10m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:08:29.512 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 3m ago 10m - - 2026-03-07T08:08:29.512 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (10m) 3m ago 10m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:08:29.698 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:08:29.698 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:08:29.698 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:08:30.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:30 vm09 bash[21220]: cluster 2026-03-07T07:08:28.936259+0000 mgr.vm07.yrfcuj (mgr.14201) 758 : cluster [DBG] pgmap v451: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:08:30.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:30 vm09 bash[21220]: cluster 2026-03-07T07:08:28.936259+0000 mgr.vm07.yrfcuj (mgr.14201) 758 : cluster [DBG] pgmap v451: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:08:30.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:30 vm09 bash[21220]: audit 2026-03-07T07:08:29.335016+0000 mgr.vm07.yrfcuj (mgr.14201) 759 : audit [DBG] from='client.16012 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:30.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:30 vm09 bash[21220]: audit 2026-03-07T07:08:29.335016+0000 mgr.vm07.yrfcuj (mgr.14201) 759 : audit [DBG] from='client.16012 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:30.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:30 vm09 bash[21220]: audit 2026-03-07T07:08:29.513052+0000 mgr.vm07.yrfcuj (mgr.14201) 760 : audit [DBG] from='client.16016 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:30.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:30 vm09 bash[21220]: audit 2026-03-07T07:08:29.513052+0000 mgr.vm07.yrfcuj (mgr.14201) 760 : audit [DBG] from='client.16016 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:30.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:30 vm09 bash[21220]: audit 2026-03-07T07:08:29.702404+0000 mon.vm07 (mon.0) 1082 : audit [DBG] from='client.? 192.168.123.107:0/3482850753' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:30.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:30 vm09 bash[21220]: audit 2026-03-07T07:08:29.702404+0000 mon.vm07 (mon.0) 1082 : audit [DBG] from='client.? 192.168.123.107:0/3482850753' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:30.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:30 vm07 bash[17031]: cluster 2026-03-07T07:08:28.936259+0000 mgr.vm07.yrfcuj (mgr.14201) 758 : cluster [DBG] pgmap v451: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:08:30.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:30 vm07 bash[17031]: cluster 2026-03-07T07:08:28.936259+0000 mgr.vm07.yrfcuj (mgr.14201) 758 : cluster [DBG] pgmap v451: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:08:30.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:30 vm07 bash[17031]: audit 2026-03-07T07:08:29.335016+0000 mgr.vm07.yrfcuj (mgr.14201) 759 : audit [DBG] from='client.16012 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:30.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:30 vm07 bash[17031]: audit 2026-03-07T07:08:29.335016+0000 mgr.vm07.yrfcuj (mgr.14201) 759 : audit [DBG] from='client.16012 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:30.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:30 vm07 bash[17031]: audit 2026-03-07T07:08:29.513052+0000 mgr.vm07.yrfcuj (mgr.14201) 760 : audit [DBG] from='client.16016 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:30.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:30 vm07 bash[17031]: audit 2026-03-07T07:08:29.513052+0000 mgr.vm07.yrfcuj (mgr.14201) 760 : audit [DBG] from='client.16016 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:30.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:30 vm07 bash[17031]: audit 2026-03-07T07:08:29.702404+0000 mon.vm07 (mon.0) 1082 : audit [DBG] from='client.? 192.168.123.107:0/3482850753' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:30.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:30 vm07 bash[17031]: audit 2026-03-07T07:08:29.702404+0000 mon.vm07 (mon.0) 1082 : audit [DBG] from='client.? 192.168.123.107:0/3482850753' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:31.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:31 vm07 bash[17031]: cluster 2026-03-07T07:08:30.936597+0000 mgr.vm07.yrfcuj (mgr.14201) 761 : cluster [DBG] pgmap v452: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:08:31.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:31 vm07 bash[17031]: cluster 2026-03-07T07:08:30.936597+0000 mgr.vm07.yrfcuj (mgr.14201) 761 : cluster [DBG] pgmap v452: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:08:32.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:31 vm09 bash[21220]: cluster 2026-03-07T07:08:30.936597+0000 mgr.vm07.yrfcuj (mgr.14201) 761 : cluster [DBG] pgmap v452: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:08:32.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:31 vm09 bash[21220]: cluster 2026-03-07T07:08:30.936597+0000 mgr.vm07.yrfcuj (mgr.14201) 761 : cluster [DBG] pgmap v452: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:08:34.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:33 vm09 bash[21220]: cluster 2026-03-07T07:08:32.936996+0000 mgr.vm07.yrfcuj (mgr.14201) 762 : cluster [DBG] pgmap v453: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:34.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:33 vm09 bash[21220]: cluster 2026-03-07T07:08:32.936996+0000 mgr.vm07.yrfcuj (mgr.14201) 762 : cluster [DBG] pgmap v453: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:34.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:34 vm07 bash[17031]: cluster 2026-03-07T07:08:32.936996+0000 mgr.vm07.yrfcuj (mgr.14201) 762 : cluster [DBG] pgmap v453: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:34.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:34 vm07 bash[17031]: cluster 2026-03-07T07:08:32.936996+0000 mgr.vm07.yrfcuj (mgr.14201) 762 : cluster [DBG] pgmap v453: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:34.874 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:08:35.027 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:08:35.027 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (5m) 4m ago 10m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:08:35.027 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 4m ago 10m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:08:35.027 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 3m ago 10m - - 2026-03-07T08:08:35.027 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (10m) 3m ago 10m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:08:35.228 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:08:35.228 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:08:35.228 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:08:36.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:36 vm09 bash[21220]: audit 2026-03-07T07:08:34.862729+0000 mgr.vm07.yrfcuj (mgr.14201) 763 : audit [DBG] from='client.16024 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:36.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:36 vm09 bash[21220]: audit 2026-03-07T07:08:34.862729+0000 mgr.vm07.yrfcuj (mgr.14201) 763 : audit [DBG] from='client.16024 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:36.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:36 vm09 bash[21220]: cluster 2026-03-07T07:08:34.937350+0000 mgr.vm07.yrfcuj (mgr.14201) 764 : cluster [DBG] pgmap v454: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:36.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:36 vm09 bash[21220]: cluster 2026-03-07T07:08:34.937350+0000 mgr.vm07.yrfcuj (mgr.14201) 764 : cluster [DBG] pgmap v454: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:36.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:36 vm09 bash[21220]: audit 2026-03-07T07:08:35.028723+0000 mgr.vm07.yrfcuj (mgr.14201) 765 : audit [DBG] from='client.16028 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:36.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:36 vm09 bash[21220]: audit 2026-03-07T07:08:35.028723+0000 mgr.vm07.yrfcuj (mgr.14201) 765 : audit [DBG] from='client.16028 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:36.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:36 vm09 bash[21220]: audit 2026-03-07T07:08:35.227793+0000 mon.vm09 (mon.1) 47 : audit [DBG] from='client.? 192.168.123.107:0/695313693' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:36.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:36 vm09 bash[21220]: audit 2026-03-07T07:08:35.227793+0000 mon.vm09 (mon.1) 47 : audit [DBG] from='client.? 192.168.123.107:0/695313693' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:36.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:36 vm07 bash[17031]: audit 2026-03-07T07:08:34.862729+0000 mgr.vm07.yrfcuj (mgr.14201) 763 : audit [DBG] from='client.16024 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:36.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:36 vm07 bash[17031]: audit 2026-03-07T07:08:34.862729+0000 mgr.vm07.yrfcuj (mgr.14201) 763 : audit [DBG] from='client.16024 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:36.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:36 vm07 bash[17031]: cluster 2026-03-07T07:08:34.937350+0000 mgr.vm07.yrfcuj (mgr.14201) 764 : cluster [DBG] pgmap v454: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:36.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:36 vm07 bash[17031]: cluster 2026-03-07T07:08:34.937350+0000 mgr.vm07.yrfcuj (mgr.14201) 764 : cluster [DBG] pgmap v454: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:36.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:36 vm07 bash[17031]: audit 2026-03-07T07:08:35.028723+0000 mgr.vm07.yrfcuj (mgr.14201) 765 : audit [DBG] from='client.16028 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:36.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:36 vm07 bash[17031]: audit 2026-03-07T07:08:35.028723+0000 mgr.vm07.yrfcuj (mgr.14201) 765 : audit [DBG] from='client.16028 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:36.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:36 vm07 bash[17031]: audit 2026-03-07T07:08:35.227793+0000 mon.vm09 (mon.1) 47 : audit [DBG] from='client.? 192.168.123.107:0/695313693' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:36.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:36 vm07 bash[17031]: audit 2026-03-07T07:08:35.227793+0000 mon.vm09 (mon.1) 47 : audit [DBG] from='client.? 192.168.123.107:0/695313693' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:37.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:37 vm09 bash[21220]: audit 2026-03-07T07:08:36.972679+0000 mon.vm07 (mon.0) 1083 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:08:37.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:37 vm09 bash[21220]: audit 2026-03-07T07:08:36.972679+0000 mon.vm07 (mon.0) 1083 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:08:37.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:37 vm07 bash[17031]: audit 2026-03-07T07:08:36.972679+0000 mon.vm07 (mon.0) 1083 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:08:37.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:37 vm07 bash[17031]: audit 2026-03-07T07:08:36.972679+0000 mon.vm07 (mon.0) 1083 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:08:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:38 vm09 bash[21220]: cluster 2026-03-07T07:08:36.937745+0000 mgr.vm07.yrfcuj (mgr.14201) 766 : cluster [DBG] pgmap v455: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:38 vm09 bash[21220]: cluster 2026-03-07T07:08:36.937745+0000 mgr.vm07.yrfcuj (mgr.14201) 766 : cluster [DBG] pgmap v455: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:38 vm09 bash[21220]: audit 2026-03-07T07:08:37.297771+0000 mon.vm07 (mon.0) 1084 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:08:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:38 vm09 bash[21220]: audit 2026-03-07T07:08:37.297771+0000 mon.vm07 (mon.0) 1084 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:08:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:38 vm09 bash[21220]: audit 2026-03-07T07:08:37.298368+0000 mon.vm07 (mon.0) 1085 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:08:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:38 vm09 bash[21220]: audit 2026-03-07T07:08:37.298368+0000 mon.vm07 (mon.0) 1085 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:08:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:38 vm09 bash[21220]: cluster 2026-03-07T07:08:37.299184+0000 mgr.vm07.yrfcuj (mgr.14201) 767 : cluster [DBG] pgmap v456: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:08:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:38 vm09 bash[21220]: cluster 2026-03-07T07:08:37.299184+0000 mgr.vm07.yrfcuj (mgr.14201) 767 : cluster [DBG] pgmap v456: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:08:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:38 vm09 bash[21220]: cluster 2026-03-07T07:08:37.299547+0000 mgr.vm07.yrfcuj (mgr.14201) 768 : cluster [DBG] pgmap v457: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:08:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:38 vm09 bash[21220]: cluster 2026-03-07T07:08:37.299547+0000 mgr.vm07.yrfcuj (mgr.14201) 768 : cluster [DBG] pgmap v457: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:08:38.366 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:38 vm09 bash[21220]: audit 2026-03-07T07:08:37.303549+0000 mon.vm07 (mon.0) 1086 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:08:38.366 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:38 vm09 bash[21220]: audit 2026-03-07T07:08:37.303549+0000 mon.vm07 (mon.0) 1086 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:08:38.366 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:38 vm09 bash[21220]: audit 2026-03-07T07:08:37.304776+0000 mon.vm07 (mon.0) 1087 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:08:38.366 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:38 vm09 bash[21220]: audit 2026-03-07T07:08:37.304776+0000 mon.vm07 (mon.0) 1087 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:08:38.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:38 vm07 bash[17031]: cluster 2026-03-07T07:08:36.937745+0000 mgr.vm07.yrfcuj (mgr.14201) 766 : cluster [DBG] pgmap v455: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:38.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:38 vm07 bash[17031]: cluster 2026-03-07T07:08:36.937745+0000 mgr.vm07.yrfcuj (mgr.14201) 766 : cluster [DBG] pgmap v455: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:38.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:38 vm07 bash[17031]: audit 2026-03-07T07:08:37.297771+0000 mon.vm07 (mon.0) 1084 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:08:38.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:38 vm07 bash[17031]: audit 2026-03-07T07:08:37.297771+0000 mon.vm07 (mon.0) 1084 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:08:38.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:38 vm07 bash[17031]: audit 2026-03-07T07:08:37.298368+0000 mon.vm07 (mon.0) 1085 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:08:38.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:38 vm07 bash[17031]: audit 2026-03-07T07:08:37.298368+0000 mon.vm07 (mon.0) 1085 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:08:38.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:38 vm07 bash[17031]: cluster 2026-03-07T07:08:37.299184+0000 mgr.vm07.yrfcuj (mgr.14201) 767 : cluster [DBG] pgmap v456: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:08:38.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:38 vm07 bash[17031]: cluster 2026-03-07T07:08:37.299184+0000 mgr.vm07.yrfcuj (mgr.14201) 767 : cluster [DBG] pgmap v456: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:08:38.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:38 vm07 bash[17031]: cluster 2026-03-07T07:08:37.299547+0000 mgr.vm07.yrfcuj (mgr.14201) 768 : cluster [DBG] pgmap v457: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:08:38.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:38 vm07 bash[17031]: cluster 2026-03-07T07:08:37.299547+0000 mgr.vm07.yrfcuj (mgr.14201) 768 : cluster [DBG] pgmap v457: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:08:38.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:38 vm07 bash[17031]: audit 2026-03-07T07:08:37.303549+0000 mon.vm07 (mon.0) 1086 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:08:38.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:38 vm07 bash[17031]: audit 2026-03-07T07:08:37.303549+0000 mon.vm07 (mon.0) 1086 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:08:38.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:38 vm07 bash[17031]: audit 2026-03-07T07:08:37.304776+0000 mon.vm07 (mon.0) 1087 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:08:38.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:38 vm07 bash[17031]: audit 2026-03-07T07:08:37.304776+0000 mon.vm07 (mon.0) 1087 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:08:40.403 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:08:40.549 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:08:40.549 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (5m) 4m ago 11m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:08:40.549 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 4m ago 11m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:08:40.549 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 4m ago 11m - - 2026-03-07T08:08:40.549 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (11m) 4m ago 11m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:08:40.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:40 vm09 bash[21220]: cluster 2026-03-07T07:08:39.299930+0000 mgr.vm07.yrfcuj (mgr.14201) 769 : cluster [DBG] pgmap v458: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:08:40.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:40 vm09 bash[21220]: cluster 2026-03-07T07:08:39.299930+0000 mgr.vm07.yrfcuj (mgr.14201) 769 : cluster [DBG] pgmap v458: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:08:40.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:40 vm09 bash[21220]: audit 2026-03-07T07:08:40.139391+0000 mon.vm07 (mon.0) 1088 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:08:40.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:40 vm09 bash[21220]: audit 2026-03-07T07:08:40.139391+0000 mon.vm07 (mon.0) 1088 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:08:40.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:40 vm07 bash[17031]: cluster 2026-03-07T07:08:39.299930+0000 mgr.vm07.yrfcuj (mgr.14201) 769 : cluster [DBG] pgmap v458: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:08:40.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:40 vm07 bash[17031]: cluster 2026-03-07T07:08:39.299930+0000 mgr.vm07.yrfcuj (mgr.14201) 769 : cluster [DBG] pgmap v458: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:08:40.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:40 vm07 bash[17031]: audit 2026-03-07T07:08:40.139391+0000 mon.vm07 (mon.0) 1088 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:08:40.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:40 vm07 bash[17031]: audit 2026-03-07T07:08:40.139391+0000 mon.vm07 (mon.0) 1088 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:08:40.735 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:08:40.735 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:08:40.735 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:08:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:41 vm09 bash[21220]: audit 2026-03-07T07:08:40.391782+0000 mgr.vm07.yrfcuj (mgr.14201) 770 : audit [DBG] from='client.25337 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:41 vm09 bash[21220]: audit 2026-03-07T07:08:40.391782+0000 mgr.vm07.yrfcuj (mgr.14201) 770 : audit [DBG] from='client.25337 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:41 vm09 bash[21220]: audit 2026-03-07T07:08:40.550538+0000 mgr.vm07.yrfcuj (mgr.14201) 771 : audit [DBG] from='client.25341 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:41 vm09 bash[21220]: audit 2026-03-07T07:08:40.550538+0000 mgr.vm07.yrfcuj (mgr.14201) 771 : audit [DBG] from='client.25341 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:41 vm09 bash[21220]: audit 2026-03-07T07:08:40.739116+0000 mon.vm07 (mon.0) 1089 : audit [DBG] from='client.? 192.168.123.107:0/3136591605' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:41 vm09 bash[21220]: audit 2026-03-07T07:08:40.739116+0000 mon.vm07 (mon.0) 1089 : audit [DBG] from='client.? 192.168.123.107:0/3136591605' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:41.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:41 vm07 bash[17031]: audit 2026-03-07T07:08:40.391782+0000 mgr.vm07.yrfcuj (mgr.14201) 770 : audit [DBG] from='client.25337 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:41.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:41 vm07 bash[17031]: audit 2026-03-07T07:08:40.391782+0000 mgr.vm07.yrfcuj (mgr.14201) 770 : audit [DBG] from='client.25337 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:41.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:41 vm07 bash[17031]: audit 2026-03-07T07:08:40.550538+0000 mgr.vm07.yrfcuj (mgr.14201) 771 : audit [DBG] from='client.25341 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:41.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:41 vm07 bash[17031]: audit 2026-03-07T07:08:40.550538+0000 mgr.vm07.yrfcuj (mgr.14201) 771 : audit [DBG] from='client.25341 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:41.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:41 vm07 bash[17031]: audit 2026-03-07T07:08:40.739116+0000 mon.vm07 (mon.0) 1089 : audit [DBG] from='client.? 192.168.123.107:0/3136591605' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:41.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:41 vm07 bash[17031]: audit 2026-03-07T07:08:40.739116+0000 mon.vm07 (mon.0) 1089 : audit [DBG] from='client.? 192.168.123.107:0/3136591605' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:42.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:42 vm09 bash[21220]: cluster 2026-03-07T07:08:41.300248+0000 mgr.vm07.yrfcuj (mgr.14201) 772 : cluster [DBG] pgmap v459: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:08:42.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:42 vm09 bash[21220]: cluster 2026-03-07T07:08:41.300248+0000 mgr.vm07.yrfcuj (mgr.14201) 772 : cluster [DBG] pgmap v459: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:08:42.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:42 vm07 bash[17031]: cluster 2026-03-07T07:08:41.300248+0000 mgr.vm07.yrfcuj (mgr.14201) 772 : cluster [DBG] pgmap v459: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:08:42.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:42 vm07 bash[17031]: cluster 2026-03-07T07:08:41.300248+0000 mgr.vm07.yrfcuj (mgr.14201) 772 : cluster [DBG] pgmap v459: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:08:44.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:44 vm07 bash[17031]: cluster 2026-03-07T07:08:43.300744+0000 mgr.vm07.yrfcuj (mgr.14201) 773 : cluster [DBG] pgmap v460: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:08:44.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:44 vm07 bash[17031]: cluster 2026-03-07T07:08:43.300744+0000 mgr.vm07.yrfcuj (mgr.14201) 773 : cluster [DBG] pgmap v460: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:08:44.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:44 vm09 bash[21220]: cluster 2026-03-07T07:08:43.300744+0000 mgr.vm07.yrfcuj (mgr.14201) 773 : cluster [DBG] pgmap v460: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:08:44.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:44 vm09 bash[21220]: cluster 2026-03-07T07:08:43.300744+0000 mgr.vm07.yrfcuj (mgr.14201) 773 : cluster [DBG] pgmap v460: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:08:45.910 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:08:46.065 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:08:46.065 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (5m) 4m ago 11m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:08:46.065 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 4m ago 11m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:08:46.065 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 4m ago 11m - - 2026-03-07T08:08:46.065 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (11m) 4m ago 11m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:08:46.250 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:08:46.250 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:08:46.250 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:08:46.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:46 vm07 bash[17031]: cluster 2026-03-07T07:08:45.301194+0000 mgr.vm07.yrfcuj (mgr.14201) 774 : cluster [DBG] pgmap v461: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:08:46.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:46 vm07 bash[17031]: cluster 2026-03-07T07:08:45.301194+0000 mgr.vm07.yrfcuj (mgr.14201) 774 : cluster [DBG] pgmap v461: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:08:46.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:46 vm07 bash[17031]: audit 2026-03-07T07:08:46.253697+0000 mon.vm07 (mon.0) 1090 : audit [DBG] from='client.? 192.168.123.107:0/157412900' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:46.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:46 vm07 bash[17031]: audit 2026-03-07T07:08:46.253697+0000 mon.vm07 (mon.0) 1090 : audit [DBG] from='client.? 192.168.123.107:0/157412900' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:46.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:46 vm09 bash[21220]: cluster 2026-03-07T07:08:45.301194+0000 mgr.vm07.yrfcuj (mgr.14201) 774 : cluster [DBG] pgmap v461: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:08:46.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:46 vm09 bash[21220]: cluster 2026-03-07T07:08:45.301194+0000 mgr.vm07.yrfcuj (mgr.14201) 774 : cluster [DBG] pgmap v461: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:08:46.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:46 vm09 bash[21220]: audit 2026-03-07T07:08:46.253697+0000 mon.vm07 (mon.0) 1090 : audit [DBG] from='client.? 192.168.123.107:0/157412900' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:46.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:46 vm09 bash[21220]: audit 2026-03-07T07:08:46.253697+0000 mon.vm07 (mon.0) 1090 : audit [DBG] from='client.? 192.168.123.107:0/157412900' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:47.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:47 vm07 bash[17031]: audit 2026-03-07T07:08:45.897717+0000 mgr.vm07.yrfcuj (mgr.14201) 775 : audit [DBG] from='client.16048 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:47.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:47 vm07 bash[17031]: audit 2026-03-07T07:08:45.897717+0000 mgr.vm07.yrfcuj (mgr.14201) 775 : audit [DBG] from='client.16048 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:47.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:47 vm07 bash[17031]: audit 2026-03-07T07:08:46.066026+0000 mgr.vm07.yrfcuj (mgr.14201) 776 : audit [DBG] from='client.16052 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:47.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:47 vm07 bash[17031]: audit 2026-03-07T07:08:46.066026+0000 mgr.vm07.yrfcuj (mgr.14201) 776 : audit [DBG] from='client.16052 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:47 vm09 bash[21220]: audit 2026-03-07T07:08:45.897717+0000 mgr.vm07.yrfcuj (mgr.14201) 775 : audit [DBG] from='client.16048 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:47 vm09 bash[21220]: audit 2026-03-07T07:08:45.897717+0000 mgr.vm07.yrfcuj (mgr.14201) 775 : audit [DBG] from='client.16048 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:47 vm09 bash[21220]: audit 2026-03-07T07:08:46.066026+0000 mgr.vm07.yrfcuj (mgr.14201) 776 : audit [DBG] from='client.16052 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:47 vm09 bash[21220]: audit 2026-03-07T07:08:46.066026+0000 mgr.vm07.yrfcuj (mgr.14201) 776 : audit [DBG] from='client.16052 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:48.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:48 vm07 bash[17031]: cluster 2026-03-07T07:08:47.301642+0000 mgr.vm07.yrfcuj (mgr.14201) 777 : cluster [DBG] pgmap v462: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 204 B/s rd, 409 B/s wr, 0 op/s 2026-03-07T08:08:48.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:48 vm07 bash[17031]: cluster 2026-03-07T07:08:47.301642+0000 mgr.vm07.yrfcuj (mgr.14201) 777 : cluster [DBG] pgmap v462: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 204 B/s rd, 409 B/s wr, 0 op/s 2026-03-07T08:08:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:48 vm09 bash[21220]: cluster 2026-03-07T07:08:47.301642+0000 mgr.vm07.yrfcuj (mgr.14201) 777 : cluster [DBG] pgmap v462: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 204 B/s rd, 409 B/s wr, 0 op/s 2026-03-07T08:08:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:48 vm09 bash[21220]: cluster 2026-03-07T07:08:47.301642+0000 mgr.vm07.yrfcuj (mgr.14201) 777 : cluster [DBG] pgmap v462: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 204 B/s rd, 409 B/s wr, 0 op/s 2026-03-07T08:08:50.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:50 vm07 bash[17031]: cluster 2026-03-07T07:08:49.302059+0000 mgr.vm07.yrfcuj (mgr.14201) 778 : cluster [DBG] pgmap v463: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:50.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:50 vm07 bash[17031]: cluster 2026-03-07T07:08:49.302059+0000 mgr.vm07.yrfcuj (mgr.14201) 778 : cluster [DBG] pgmap v463: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:50 vm09 bash[21220]: cluster 2026-03-07T07:08:49.302059+0000 mgr.vm07.yrfcuj (mgr.14201) 778 : cluster [DBG] pgmap v463: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:50 vm09 bash[21220]: cluster 2026-03-07T07:08:49.302059+0000 mgr.vm07.yrfcuj (mgr.14201) 778 : cluster [DBG] pgmap v463: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:51.426 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:08:51.573 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:08:51.573 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (5m) 4m ago 11m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:08:51.573 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 4m ago 11m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:08:51.573 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 4m ago 11m - - 2026-03-07T08:08:51.573 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (11m) 4m ago 11m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:08:51.756 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:08:51.756 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:08:51.756 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:08:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:52 vm09 bash[21220]: cluster 2026-03-07T07:08:51.302443+0000 mgr.vm07.yrfcuj (mgr.14201) 779 : cluster [DBG] pgmap v464: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:52 vm09 bash[21220]: cluster 2026-03-07T07:08:51.302443+0000 mgr.vm07.yrfcuj (mgr.14201) 779 : cluster [DBG] pgmap v464: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:52 vm09 bash[21220]: audit 2026-03-07T07:08:51.415600+0000 mgr.vm07.yrfcuj (mgr.14201) 780 : audit [DBG] from='client.25355 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:52 vm09 bash[21220]: audit 2026-03-07T07:08:51.415600+0000 mgr.vm07.yrfcuj (mgr.14201) 780 : audit [DBG] from='client.25355 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:52 vm09 bash[21220]: audit 2026-03-07T07:08:51.574986+0000 mgr.vm07.yrfcuj (mgr.14201) 781 : audit [DBG] from='client.25359 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:52 vm09 bash[21220]: audit 2026-03-07T07:08:51.574986+0000 mgr.vm07.yrfcuj (mgr.14201) 781 : audit [DBG] from='client.25359 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:52 vm09 bash[21220]: audit 2026-03-07T07:08:51.760204+0000 mon.vm07 (mon.0) 1091 : audit [DBG] from='client.? 192.168.123.107:0/182172749' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:52 vm09 bash[21220]: audit 2026-03-07T07:08:51.760204+0000 mon.vm07 (mon.0) 1091 : audit [DBG] from='client.? 192.168.123.107:0/182172749' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:52.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:52 vm07 bash[17031]: cluster 2026-03-07T07:08:51.302443+0000 mgr.vm07.yrfcuj (mgr.14201) 779 : cluster [DBG] pgmap v464: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:52.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:52 vm07 bash[17031]: cluster 2026-03-07T07:08:51.302443+0000 mgr.vm07.yrfcuj (mgr.14201) 779 : cluster [DBG] pgmap v464: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:52.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:52 vm07 bash[17031]: audit 2026-03-07T07:08:51.415600+0000 mgr.vm07.yrfcuj (mgr.14201) 780 : audit [DBG] from='client.25355 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:52.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:52 vm07 bash[17031]: audit 2026-03-07T07:08:51.415600+0000 mgr.vm07.yrfcuj (mgr.14201) 780 : audit [DBG] from='client.25355 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:52.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:52 vm07 bash[17031]: audit 2026-03-07T07:08:51.574986+0000 mgr.vm07.yrfcuj (mgr.14201) 781 : audit [DBG] from='client.25359 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:52.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:52 vm07 bash[17031]: audit 2026-03-07T07:08:51.574986+0000 mgr.vm07.yrfcuj (mgr.14201) 781 : audit [DBG] from='client.25359 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:52.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:52 vm07 bash[17031]: audit 2026-03-07T07:08:51.760204+0000 mon.vm07 (mon.0) 1091 : audit [DBG] from='client.? 192.168.123.107:0/182172749' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:52.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:52 vm07 bash[17031]: audit 2026-03-07T07:08:51.760204+0000 mon.vm07 (mon.0) 1091 : audit [DBG] from='client.? 192.168.123.107:0/182172749' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:54.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:54 vm09 bash[21220]: cluster 2026-03-07T07:08:53.302932+0000 mgr.vm07.yrfcuj (mgr.14201) 782 : cluster [DBG] pgmap v465: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:54.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:54 vm09 bash[21220]: cluster 2026-03-07T07:08:53.302932+0000 mgr.vm07.yrfcuj (mgr.14201) 782 : cluster [DBG] pgmap v465: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:54.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:54 vm07 bash[17031]: cluster 2026-03-07T07:08:53.302932+0000 mgr.vm07.yrfcuj (mgr.14201) 782 : cluster [DBG] pgmap v465: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:54.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:54 vm07 bash[17031]: cluster 2026-03-07T07:08:53.302932+0000 mgr.vm07.yrfcuj (mgr.14201) 782 : cluster [DBG] pgmap v465: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:08:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:55 vm09 bash[21220]: audit 2026-03-07T07:08:55.139593+0000 mon.vm07 (mon.0) 1092 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:08:55.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:55 vm09 bash[21220]: audit 2026-03-07T07:08:55.139593+0000 mon.vm07 (mon.0) 1092 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:08:55.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:55 vm07 bash[17031]: audit 2026-03-07T07:08:55.139593+0000 mon.vm07 (mon.0) 1092 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:08:55.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:55 vm07 bash[17031]: audit 2026-03-07T07:08:55.139593+0000 mon.vm07 (mon.0) 1092 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:08:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:56 vm09 bash[21220]: cluster 2026-03-07T07:08:55.303300+0000 mgr.vm07.yrfcuj (mgr.14201) 783 : cluster [DBG] pgmap v466: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:56 vm09 bash[21220]: cluster 2026-03-07T07:08:55.303300+0000 mgr.vm07.yrfcuj (mgr.14201) 783 : cluster [DBG] pgmap v466: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:56.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:56 vm07 bash[17031]: cluster 2026-03-07T07:08:55.303300+0000 mgr.vm07.yrfcuj (mgr.14201) 783 : cluster [DBG] pgmap v466: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:56.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:56 vm07 bash[17031]: cluster 2026-03-07T07:08:55.303300+0000 mgr.vm07.yrfcuj (mgr.14201) 783 : cluster [DBG] pgmap v466: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:56.928 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:08:57.086 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:08:57.086 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (5m) 4m ago 11m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:08:57.086 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 4m ago 11m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:08:57.086 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 4m ago 11m - - 2026-03-07T08:08:57.086 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (11m) 4m ago 11m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:08:57.279 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:08:57.279 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:08:57.279 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:08:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:57 vm09 bash[21220]: audit 2026-03-07T07:08:57.283183+0000 mon.vm07 (mon.0) 1093 : audit [DBG] from='client.? 192.168.123.107:0/1391556741' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:57.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:57 vm09 bash[21220]: audit 2026-03-07T07:08:57.283183+0000 mon.vm07 (mon.0) 1093 : audit [DBG] from='client.? 192.168.123.107:0/1391556741' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:57.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:57 vm07 bash[17031]: audit 2026-03-07T07:08:57.283183+0000 mon.vm07 (mon.0) 1093 : audit [DBG] from='client.? 192.168.123.107:0/1391556741' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:57.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:57 vm07 bash[17031]: audit 2026-03-07T07:08:57.283183+0000 mon.vm07 (mon.0) 1093 : audit [DBG] from='client.? 192.168.123.107:0/1391556741' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:08:58.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:58 vm09 bash[21220]: audit 2026-03-07T07:08:56.917732+0000 mgr.vm07.yrfcuj (mgr.14201) 784 : audit [DBG] from='client.16072 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:58.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:58 vm09 bash[21220]: audit 2026-03-07T07:08:56.917732+0000 mgr.vm07.yrfcuj (mgr.14201) 784 : audit [DBG] from='client.16072 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:58.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:58 vm09 bash[21220]: audit 2026-03-07T07:08:57.087541+0000 mgr.vm07.yrfcuj (mgr.14201) 785 : audit [DBG] from='client.16076 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:58.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:58 vm09 bash[21220]: audit 2026-03-07T07:08:57.087541+0000 mgr.vm07.yrfcuj (mgr.14201) 785 : audit [DBG] from='client.16076 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:58.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:58 vm09 bash[21220]: cluster 2026-03-07T07:08:57.303701+0000 mgr.vm07.yrfcuj (mgr.14201) 786 : cluster [DBG] pgmap v467: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:58.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:08:58 vm09 bash[21220]: cluster 2026-03-07T07:08:57.303701+0000 mgr.vm07.yrfcuj (mgr.14201) 786 : cluster [DBG] pgmap v467: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:58.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:58 vm07 bash[17031]: audit 2026-03-07T07:08:56.917732+0000 mgr.vm07.yrfcuj (mgr.14201) 784 : audit [DBG] from='client.16072 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:58.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:58 vm07 bash[17031]: audit 2026-03-07T07:08:56.917732+0000 mgr.vm07.yrfcuj (mgr.14201) 784 : audit [DBG] from='client.16072 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:58.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:58 vm07 bash[17031]: audit 2026-03-07T07:08:57.087541+0000 mgr.vm07.yrfcuj (mgr.14201) 785 : audit [DBG] from='client.16076 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:58.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:58 vm07 bash[17031]: audit 2026-03-07T07:08:57.087541+0000 mgr.vm07.yrfcuj (mgr.14201) 785 : audit [DBG] from='client.16076 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:08:58.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:58 vm07 bash[17031]: cluster 2026-03-07T07:08:57.303701+0000 mgr.vm07.yrfcuj (mgr.14201) 786 : cluster [DBG] pgmap v467: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:08:58.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:08:58 vm07 bash[17031]: cluster 2026-03-07T07:08:57.303701+0000 mgr.vm07.yrfcuj (mgr.14201) 786 : cluster [DBG] pgmap v467: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:00.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:00 vm07 bash[17031]: cluster 2026-03-07T07:08:59.304136+0000 mgr.vm07.yrfcuj (mgr.14201) 787 : cluster [DBG] pgmap v468: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:09:00.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:00 vm07 bash[17031]: cluster 2026-03-07T07:08:59.304136+0000 mgr.vm07.yrfcuj (mgr.14201) 787 : cluster [DBG] pgmap v468: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:09:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:00 vm09 bash[21220]: cluster 2026-03-07T07:08:59.304136+0000 mgr.vm07.yrfcuj (mgr.14201) 787 : cluster [DBG] pgmap v468: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:09:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:00 vm09 bash[21220]: cluster 2026-03-07T07:08:59.304136+0000 mgr.vm07.yrfcuj (mgr.14201) 787 : cluster [DBG] pgmap v468: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:09:02.451 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:09:02.598 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:09:02.598 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (5m) 4m ago 11m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:09:02.598 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (4m) 4m ago 11m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:09:02.598 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 4m ago 11m - - 2026-03-07T08:09:02.598 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (11m) 4m ago 11m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:09:02.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:02 vm07 bash[17031]: cluster 2026-03-07T07:09:01.304521+0000 mgr.vm07.yrfcuj (mgr.14201) 788 : cluster [DBG] pgmap v469: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:09:02.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:02 vm07 bash[17031]: cluster 2026-03-07T07:09:01.304521+0000 mgr.vm07.yrfcuj (mgr.14201) 788 : cluster [DBG] pgmap v469: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:09:02.781 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:09:02.781 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:09:02.781 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:09:02.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:02 vm09 bash[21220]: cluster 2026-03-07T07:09:01.304521+0000 mgr.vm07.yrfcuj (mgr.14201) 788 : cluster [DBG] pgmap v469: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:09:02.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:02 vm09 bash[21220]: cluster 2026-03-07T07:09:01.304521+0000 mgr.vm07.yrfcuj (mgr.14201) 788 : cluster [DBG] pgmap v469: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:09:03.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:03 vm07 bash[17031]: audit 2026-03-07T07:09:02.441324+0000 mgr.vm07.yrfcuj (mgr.14201) 789 : audit [DBG] from='client.16084 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:03.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:03 vm07 bash[17031]: audit 2026-03-07T07:09:02.441324+0000 mgr.vm07.yrfcuj (mgr.14201) 789 : audit [DBG] from='client.16084 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:03.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:03 vm07 bash[17031]: audit 2026-03-07T07:09:02.785577+0000 mon.vm07 (mon.0) 1094 : audit [DBG] from='client.? 192.168.123.107:0/2460554240' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:03.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:03 vm07 bash[17031]: audit 2026-03-07T07:09:02.785577+0000 mon.vm07 (mon.0) 1094 : audit [DBG] from='client.? 192.168.123.107:0/2460554240' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:03.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:03 vm09 bash[21220]: audit 2026-03-07T07:09:02.441324+0000 mgr.vm07.yrfcuj (mgr.14201) 789 : audit [DBG] from='client.16084 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:03.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:03 vm09 bash[21220]: audit 2026-03-07T07:09:02.441324+0000 mgr.vm07.yrfcuj (mgr.14201) 789 : audit [DBG] from='client.16084 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:03.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:03 vm09 bash[21220]: audit 2026-03-07T07:09:02.785577+0000 mon.vm07 (mon.0) 1094 : audit [DBG] from='client.? 192.168.123.107:0/2460554240' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:03.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:03 vm09 bash[21220]: audit 2026-03-07T07:09:02.785577+0000 mon.vm07 (mon.0) 1094 : audit [DBG] from='client.? 192.168.123.107:0/2460554240' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:04.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:04 vm09 bash[21220]: audit 2026-03-07T07:09:02.600055+0000 mgr.vm07.yrfcuj (mgr.14201) 790 : audit [DBG] from='client.16088 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:04.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:04 vm09 bash[21220]: audit 2026-03-07T07:09:02.600055+0000 mgr.vm07.yrfcuj (mgr.14201) 790 : audit [DBG] from='client.16088 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:04.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:04 vm09 bash[21220]: cluster 2026-03-07T07:09:03.304989+0000 mgr.vm07.yrfcuj (mgr.14201) 791 : cluster [DBG] pgmap v470: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:04.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:04 vm09 bash[21220]: cluster 2026-03-07T07:09:03.304989+0000 mgr.vm07.yrfcuj (mgr.14201) 791 : cluster [DBG] pgmap v470: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:04.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:04 vm07 bash[17031]: audit 2026-03-07T07:09:02.600055+0000 mgr.vm07.yrfcuj (mgr.14201) 790 : audit [DBG] from='client.16088 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:04.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:04 vm07 bash[17031]: audit 2026-03-07T07:09:02.600055+0000 mgr.vm07.yrfcuj (mgr.14201) 790 : audit [DBG] from='client.16088 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:04.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:04 vm07 bash[17031]: cluster 2026-03-07T07:09:03.304989+0000 mgr.vm07.yrfcuj (mgr.14201) 791 : cluster [DBG] pgmap v470: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:04.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:04 vm07 bash[17031]: cluster 2026-03-07T07:09:03.304989+0000 mgr.vm07.yrfcuj (mgr.14201) 791 : cluster [DBG] pgmap v470: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:06.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:06 vm07 bash[17031]: cluster 2026-03-07T07:09:05.305352+0000 mgr.vm07.yrfcuj (mgr.14201) 792 : cluster [DBG] pgmap v471: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:06.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:06 vm07 bash[17031]: cluster 2026-03-07T07:09:05.305352+0000 mgr.vm07.yrfcuj (mgr.14201) 792 : cluster [DBG] pgmap v471: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:06 vm09 bash[21220]: cluster 2026-03-07T07:09:05.305352+0000 mgr.vm07.yrfcuj (mgr.14201) 792 : cluster [DBG] pgmap v471: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:06 vm09 bash[21220]: cluster 2026-03-07T07:09:05.305352+0000 mgr.vm07.yrfcuj (mgr.14201) 792 : cluster [DBG] pgmap v471: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:07.954 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:09:08.111 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:09:08.111 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (5m) 4m ago 11m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:09:08.111 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 4m ago 11m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:09:08.111 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 4m ago 11m - - 2026-03-07T08:09:08.111 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (11m) 4m ago 11m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:09:08.297 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:09:08.297 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:09:08.297 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:09:08.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:08 vm09 bash[21220]: cluster 2026-03-07T07:09:07.305767+0000 mgr.vm07.yrfcuj (mgr.14201) 793 : cluster [DBG] pgmap v472: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:08.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:08 vm09 bash[21220]: cluster 2026-03-07T07:09:07.305767+0000 mgr.vm07.yrfcuj (mgr.14201) 793 : cluster [DBG] pgmap v472: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:08.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:08 vm09 bash[21220]: audit 2026-03-07T07:09:08.296141+0000 mon.vm09 (mon.1) 48 : audit [DBG] from='client.? 192.168.123.107:0/3596591296' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:08.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:08 vm09 bash[21220]: audit 2026-03-07T07:09:08.296141+0000 mon.vm09 (mon.1) 48 : audit [DBG] from='client.? 192.168.123.107:0/3596591296' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:08.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:08 vm07 bash[17031]: cluster 2026-03-07T07:09:07.305767+0000 mgr.vm07.yrfcuj (mgr.14201) 793 : cluster [DBG] pgmap v472: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:08.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:08 vm07 bash[17031]: cluster 2026-03-07T07:09:07.305767+0000 mgr.vm07.yrfcuj (mgr.14201) 793 : cluster [DBG] pgmap v472: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:08.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:08 vm07 bash[17031]: audit 2026-03-07T07:09:08.296141+0000 mon.vm09 (mon.1) 48 : audit [DBG] from='client.? 192.168.123.107:0/3596591296' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:08.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:08 vm07 bash[17031]: audit 2026-03-07T07:09:08.296141+0000 mon.vm09 (mon.1) 48 : audit [DBG] from='client.? 192.168.123.107:0/3596591296' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:09.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:09 vm09 bash[21220]: audit 2026-03-07T07:09:07.942130+0000 mgr.vm07.yrfcuj (mgr.14201) 794 : audit [DBG] from='client.16096 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:09.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:09 vm09 bash[21220]: audit 2026-03-07T07:09:07.942130+0000 mgr.vm07.yrfcuj (mgr.14201) 794 : audit [DBG] from='client.16096 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:09.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:09 vm09 bash[21220]: audit 2026-03-07T07:09:08.111670+0000 mgr.vm07.yrfcuj (mgr.14201) 795 : audit [DBG] from='client.16100 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:09.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:09 vm09 bash[21220]: audit 2026-03-07T07:09:08.111670+0000 mgr.vm07.yrfcuj (mgr.14201) 795 : audit [DBG] from='client.16100 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:09.879 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:09 vm07 bash[17031]: audit 2026-03-07T07:09:07.942130+0000 mgr.vm07.yrfcuj (mgr.14201) 794 : audit [DBG] from='client.16096 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:09.879 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:09 vm07 bash[17031]: audit 2026-03-07T07:09:07.942130+0000 mgr.vm07.yrfcuj (mgr.14201) 794 : audit [DBG] from='client.16096 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:09.879 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:09 vm07 bash[17031]: audit 2026-03-07T07:09:08.111670+0000 mgr.vm07.yrfcuj (mgr.14201) 795 : audit [DBG] from='client.16100 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:09.879 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:09 vm07 bash[17031]: audit 2026-03-07T07:09:08.111670+0000 mgr.vm07.yrfcuj (mgr.14201) 795 : audit [DBG] from='client.16100 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:10 vm09 bash[21220]: cluster 2026-03-07T07:09:09.306332+0000 mgr.vm07.yrfcuj (mgr.14201) 796 : cluster [DBG] pgmap v473: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:10 vm09 bash[21220]: cluster 2026-03-07T07:09:09.306332+0000 mgr.vm07.yrfcuj (mgr.14201) 796 : cluster [DBG] pgmap v473: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:10 vm09 bash[21220]: audit 2026-03-07T07:09:10.140123+0000 mon.vm07 (mon.0) 1095 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:09:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:10 vm09 bash[21220]: audit 2026-03-07T07:09:10.140123+0000 mon.vm07 (mon.0) 1095 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:09:10.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:10 vm07 bash[17031]: cluster 2026-03-07T07:09:09.306332+0000 mgr.vm07.yrfcuj (mgr.14201) 796 : cluster [DBG] pgmap v473: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:10.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:10 vm07 bash[17031]: cluster 2026-03-07T07:09:09.306332+0000 mgr.vm07.yrfcuj (mgr.14201) 796 : cluster [DBG] pgmap v473: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:10.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:10 vm07 bash[17031]: audit 2026-03-07T07:09:10.140123+0000 mon.vm07 (mon.0) 1095 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:09:10.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:10 vm07 bash[17031]: audit 2026-03-07T07:09:10.140123+0000 mon.vm07 (mon.0) 1095 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:09:12.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:12 vm09 bash[21220]: cluster 2026-03-07T07:09:11.306711+0000 mgr.vm07.yrfcuj (mgr.14201) 797 : cluster [DBG] pgmap v474: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:12.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:12 vm09 bash[21220]: cluster 2026-03-07T07:09:11.306711+0000 mgr.vm07.yrfcuj (mgr.14201) 797 : cluster [DBG] pgmap v474: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:12.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:12 vm07 bash[17031]: cluster 2026-03-07T07:09:11.306711+0000 mgr.vm07.yrfcuj (mgr.14201) 797 : cluster [DBG] pgmap v474: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:12.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:12 vm07 bash[17031]: cluster 2026-03-07T07:09:11.306711+0000 mgr.vm07.yrfcuj (mgr.14201) 797 : cluster [DBG] pgmap v474: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:13.485 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to stop 2026-03-07T08:09:13.639 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:09:13.639 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (5m) 5m ago 11m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:09:13.639 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 5m ago 11m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:09:13.640 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 4m ago 11m - - 2026-03-07T08:09:13.640 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (11m) 4m ago 11m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:09:13.834 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:09:13.834 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:09:13.834 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:09:14.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:14 vm09 bash[21220]: cluster 2026-03-07T07:09:13.307261+0000 mgr.vm07.yrfcuj (mgr.14201) 798 : cluster [DBG] pgmap v475: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:14.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:14 vm09 bash[21220]: cluster 2026-03-07T07:09:13.307261+0000 mgr.vm07.yrfcuj (mgr.14201) 798 : cluster [DBG] pgmap v475: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:14.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:14 vm09 bash[21220]: audit 2026-03-07T07:09:13.473267+0000 mgr.vm07.yrfcuj (mgr.14201) 799 : audit [DBG] from='client.16108 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:14.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:14 vm09 bash[21220]: audit 2026-03-07T07:09:13.473267+0000 mgr.vm07.yrfcuj (mgr.14201) 799 : audit [DBG] from='client.16108 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:14.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:14 vm09 bash[21220]: audit 2026-03-07T07:09:13.837484+0000 mon.vm07 (mon.0) 1096 : audit [DBG] from='client.? 192.168.123.107:0/3558758322' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:14.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:14 vm09 bash[21220]: audit 2026-03-07T07:09:13.837484+0000 mon.vm07 (mon.0) 1096 : audit [DBG] from='client.? 192.168.123.107:0/3558758322' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:14.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:14 vm07 bash[17031]: cluster 2026-03-07T07:09:13.307261+0000 mgr.vm07.yrfcuj (mgr.14201) 798 : cluster [DBG] pgmap v475: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:14.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:14 vm07 bash[17031]: cluster 2026-03-07T07:09:13.307261+0000 mgr.vm07.yrfcuj (mgr.14201) 798 : cluster [DBG] pgmap v475: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:14.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:14 vm07 bash[17031]: audit 2026-03-07T07:09:13.473267+0000 mgr.vm07.yrfcuj (mgr.14201) 799 : audit [DBG] from='client.16108 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:14.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:14 vm07 bash[17031]: audit 2026-03-07T07:09:13.473267+0000 mgr.vm07.yrfcuj (mgr.14201) 799 : audit [DBG] from='client.16108 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:14.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:14 vm07 bash[17031]: audit 2026-03-07T07:09:13.837484+0000 mon.vm07 (mon.0) 1096 : audit [DBG] from='client.? 192.168.123.107:0/3558758322' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:14.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:14 vm07 bash[17031]: audit 2026-03-07T07:09:13.837484+0000 mon.vm07 (mon.0) 1096 : audit [DBG] from='client.? 192.168.123.107:0/3558758322' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:15.287 INFO:teuthology.orchestra.run.vm07.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-07T08:09:15.291 INFO:teuthology.orchestra.run.vm07.stderr: Dload Upload Total Spent Left Speed 2026-03-07T08:09:15.299 INFO:teuthology.orchestra.run.vm07.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 14384 0 --:--:-- --:--:-- --:--:-- 14384 2026-03-07T08:09:15.491 INFO:teuthology.orchestra.run.vm07.stdout:anonymousScheduled to start rgw.foo.vm09.fgzozy on host 'vm09' 2026-03-07T08:09:15.690 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to start 2026-03-07T08:09:15.848 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:09:15.848 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (5m) 5m ago 11m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:09:15.848 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 5m ago 11m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:09:15.848 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 4m ago 11m - - 2026-03-07T08:09:15.848 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (11m) 4m ago 11m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:09:15.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:15 vm09 bash[21220]: audit 2026-03-07T07:09:13.640652+0000 mgr.vm07.yrfcuj (mgr.14201) 800 : audit [DBG] from='client.16112 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:15.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:15 vm09 bash[21220]: audit 2026-03-07T07:09:13.640652+0000 mgr.vm07.yrfcuj (mgr.14201) 800 : audit [DBG] from='client.16112 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:15.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:15 vm07 bash[17031]: audit 2026-03-07T07:09:13.640652+0000 mgr.vm07.yrfcuj (mgr.14201) 800 : audit [DBG] from='client.16112 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:15.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:15 vm07 bash[17031]: audit 2026-03-07T07:09:13.640652+0000 mgr.vm07.yrfcuj (mgr.14201) 800 : audit [DBG] from='client.16112 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:16.041 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:09:16.041 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:09:16.041 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:09:16.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: cluster 2026-03-07T07:09:15.307673+0000 mgr.vm07.yrfcuj (mgr.14201) 801 : cluster [DBG] pgmap v476: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:16.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: cluster 2026-03-07T07:09:15.307673+0000 mgr.vm07.yrfcuj (mgr.14201) 801 : cluster [DBG] pgmap v476: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: audit 2026-03-07T07:09:15.480577+0000 mgr.vm07.yrfcuj (mgr.14201) 802 : audit [DBG] from='client.16120 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "rgw.foo.vm09.fgzozy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: audit 2026-03-07T07:09:15.480577+0000 mgr.vm07.yrfcuj (mgr.14201) 802 : audit [DBG] from='client.16120 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "rgw.foo.vm09.fgzozy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: cephadm 2026-03-07T07:09:15.480997+0000 mgr.vm07.yrfcuj (mgr.14201) 803 : cephadm [INF] Schedule start daemon rgw.foo.vm09.fgzozy 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: cephadm 2026-03-07T07:09:15.480997+0000 mgr.vm07.yrfcuj (mgr.14201) 803 : cephadm [INF] Schedule start daemon rgw.foo.vm09.fgzozy 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: audit 2026-03-07T07:09:15.486824+0000 mon.vm07 (mon.0) 1097 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: audit 2026-03-07T07:09:15.486824+0000 mon.vm07 (mon.0) 1097 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: audit 2026-03-07T07:09:15.493328+0000 mon.vm07 (mon.0) 1098 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: audit 2026-03-07T07:09:15.493328+0000 mon.vm07 (mon.0) 1098 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: audit 2026-03-07T07:09:15.494360+0000 mon.vm07 (mon.0) 1099 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: audit 2026-03-07T07:09:15.494360+0000 mon.vm07 (mon.0) 1099 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: audit 2026-03-07T07:09:15.496627+0000 mon.vm07 (mon.0) 1100 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: audit 2026-03-07T07:09:15.496627+0000 mon.vm07 (mon.0) 1100 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: audit 2026-03-07T07:09:15.497401+0000 mon.vm07 (mon.0) 1101 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: audit 2026-03-07T07:09:15.497401+0000 mon.vm07 (mon.0) 1101 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: cluster 2026-03-07T07:09:15.499898+0000 mgr.vm07.yrfcuj (mgr.14201) 804 : cluster [DBG] pgmap v477: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 100 B/s rd, 200 B/s wr, 0 op/s 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: cluster 2026-03-07T07:09:15.499898+0000 mgr.vm07.yrfcuj (mgr.14201) 804 : cluster [DBG] pgmap v477: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 100 B/s rd, 200 B/s wr, 0 op/s 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: audit 2026-03-07T07:09:15.504249+0000 mon.vm07 (mon.0) 1102 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: audit 2026-03-07T07:09:15.504249+0000 mon.vm07 (mon.0) 1102 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: audit 2026-03-07T07:09:15.505903+0000 mon.vm07 (mon.0) 1103 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: audit 2026-03-07T07:09:15.505903+0000 mon.vm07 (mon.0) 1103 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: audit 2026-03-07T07:09:16.045041+0000 mon.vm07 (mon.0) 1104 : audit [DBG] from='client.? 192.168.123.107:0/3279708533' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:16.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:16 vm09 bash[21220]: audit 2026-03-07T07:09:16.045041+0000 mon.vm07 (mon.0) 1104 : audit [DBG] from='client.? 192.168.123.107:0/3279708533' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: cluster 2026-03-07T07:09:15.307673+0000 mgr.vm07.yrfcuj (mgr.14201) 801 : cluster [DBG] pgmap v476: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: cluster 2026-03-07T07:09:15.307673+0000 mgr.vm07.yrfcuj (mgr.14201) 801 : cluster [DBG] pgmap v476: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: audit 2026-03-07T07:09:15.480577+0000 mgr.vm07.yrfcuj (mgr.14201) 802 : audit [DBG] from='client.16120 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "rgw.foo.vm09.fgzozy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: audit 2026-03-07T07:09:15.480577+0000 mgr.vm07.yrfcuj (mgr.14201) 802 : audit [DBG] from='client.16120 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "rgw.foo.vm09.fgzozy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: cephadm 2026-03-07T07:09:15.480997+0000 mgr.vm07.yrfcuj (mgr.14201) 803 : cephadm [INF] Schedule start daemon rgw.foo.vm09.fgzozy 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: cephadm 2026-03-07T07:09:15.480997+0000 mgr.vm07.yrfcuj (mgr.14201) 803 : cephadm [INF] Schedule start daemon rgw.foo.vm09.fgzozy 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: audit 2026-03-07T07:09:15.486824+0000 mon.vm07 (mon.0) 1097 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: audit 2026-03-07T07:09:15.486824+0000 mon.vm07 (mon.0) 1097 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: audit 2026-03-07T07:09:15.493328+0000 mon.vm07 (mon.0) 1098 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: audit 2026-03-07T07:09:15.493328+0000 mon.vm07 (mon.0) 1098 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: audit 2026-03-07T07:09:15.494360+0000 mon.vm07 (mon.0) 1099 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: audit 2026-03-07T07:09:15.494360+0000 mon.vm07 (mon.0) 1099 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: audit 2026-03-07T07:09:15.496627+0000 mon.vm07 (mon.0) 1100 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: audit 2026-03-07T07:09:15.496627+0000 mon.vm07 (mon.0) 1100 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: audit 2026-03-07T07:09:15.497401+0000 mon.vm07 (mon.0) 1101 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: audit 2026-03-07T07:09:15.497401+0000 mon.vm07 (mon.0) 1101 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: cluster 2026-03-07T07:09:15.499898+0000 mgr.vm07.yrfcuj (mgr.14201) 804 : cluster [DBG] pgmap v477: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 100 B/s rd, 200 B/s wr, 0 op/s 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: cluster 2026-03-07T07:09:15.499898+0000 mgr.vm07.yrfcuj (mgr.14201) 804 : cluster [DBG] pgmap v477: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 100 B/s rd, 200 B/s wr, 0 op/s 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: audit 2026-03-07T07:09:15.504249+0000 mon.vm07 (mon.0) 1102 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: audit 2026-03-07T07:09:15.504249+0000 mon.vm07 (mon.0) 1102 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: audit 2026-03-07T07:09:15.505903+0000 mon.vm07 (mon.0) 1103 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: audit 2026-03-07T07:09:15.505903+0000 mon.vm07 (mon.0) 1103 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: audit 2026-03-07T07:09:16.045041+0000 mon.vm07 (mon.0) 1104 : audit [DBG] from='client.? 192.168.123.107:0/3279708533' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:16.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:16 vm07 bash[17031]: audit 2026-03-07T07:09:16.045041+0000 mon.vm07 (mon.0) 1104 : audit [DBG] from='client.? 192.168.123.107:0/3279708533' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:17.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:17 vm09 bash[21220]: audit 2026-03-07T07:09:15.677636+0000 mgr.vm07.yrfcuj (mgr.14201) 805 : audit [DBG] from='client.16124 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:17.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:17 vm09 bash[21220]: audit 2026-03-07T07:09:15.677636+0000 mgr.vm07.yrfcuj (mgr.14201) 805 : audit [DBG] from='client.16124 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:17.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:17 vm09 bash[21220]: audit 2026-03-07T07:09:15.848965+0000 mgr.vm07.yrfcuj (mgr.14201) 806 : audit [DBG] from='client.16128 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:17.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:17 vm09 bash[21220]: audit 2026-03-07T07:09:15.848965+0000 mgr.vm07.yrfcuj (mgr.14201) 806 : audit [DBG] from='client.16128 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:17.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:17 vm07 bash[17031]: audit 2026-03-07T07:09:15.677636+0000 mgr.vm07.yrfcuj (mgr.14201) 805 : audit [DBG] from='client.16124 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:17.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:17 vm07 bash[17031]: audit 2026-03-07T07:09:15.677636+0000 mgr.vm07.yrfcuj (mgr.14201) 805 : audit [DBG] from='client.16124 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:17.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:17 vm07 bash[17031]: audit 2026-03-07T07:09:15.848965+0000 mgr.vm07.yrfcuj (mgr.14201) 806 : audit [DBG] from='client.16128 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:17.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:17 vm07 bash[17031]: audit 2026-03-07T07:09:15.848965+0000 mgr.vm07.yrfcuj (mgr.14201) 806 : audit [DBG] from='client.16128 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:18 vm09 bash[21220]: cluster 2026-03-07T07:09:17.500287+0000 mgr.vm07.yrfcuj (mgr.14201) 807 : cluster [DBG] pgmap v478: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 200 B/s rd, 401 B/s wr, 0 op/s 2026-03-07T08:09:18.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:18 vm09 bash[21220]: cluster 2026-03-07T07:09:17.500287+0000 mgr.vm07.yrfcuj (mgr.14201) 807 : cluster [DBG] pgmap v478: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 200 B/s rd, 401 B/s wr, 0 op/s 2026-03-07T08:09:18.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:18 vm07 bash[17031]: cluster 2026-03-07T07:09:17.500287+0000 mgr.vm07.yrfcuj (mgr.14201) 807 : cluster [DBG] pgmap v478: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 200 B/s rd, 401 B/s wr, 0 op/s 2026-03-07T08:09:18.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:18 vm07 bash[17031]: cluster 2026-03-07T07:09:17.500287+0000 mgr.vm07.yrfcuj (mgr.14201) 807 : cluster [DBG] pgmap v478: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 200 B/s rd, 401 B/s wr, 0 op/s 2026-03-07T08:09:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:20 vm09 bash[21220]: cluster 2026-03-07T07:09:19.500737+0000 mgr.vm07.yrfcuj (mgr.14201) 808 : cluster [DBG] pgmap v479: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 200 B/s rd, 401 B/s wr, 0 op/s 2026-03-07T08:09:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:20 vm09 bash[21220]: cluster 2026-03-07T07:09:19.500737+0000 mgr.vm07.yrfcuj (mgr.14201) 808 : cluster [DBG] pgmap v479: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 200 B/s rd, 401 B/s wr, 0 op/s 2026-03-07T08:09:20.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:20 vm07 bash[17031]: cluster 2026-03-07T07:09:19.500737+0000 mgr.vm07.yrfcuj (mgr.14201) 808 : cluster [DBG] pgmap v479: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 200 B/s rd, 401 B/s wr, 0 op/s 2026-03-07T08:09:20.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:20 vm07 bash[17031]: cluster 2026-03-07T07:09:19.500737+0000 mgr.vm07.yrfcuj (mgr.14201) 808 : cluster [DBG] pgmap v479: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 200 B/s rd, 401 B/s wr, 0 op/s 2026-03-07T08:09:21.235 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to start 2026-03-07T08:09:21.402 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:09:21.403 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (5m) 5m ago 11m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:09:21.403 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 5m ago 11m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:09:21.403 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 4m ago 11m - - 2026-03-07T08:09:21.403 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (11m) 4m ago 11m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:09:21.601 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:09:21.602 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:09:21.602 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:09:22.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:22 vm09 bash[21220]: audit 2026-03-07T07:09:21.223255+0000 mgr.vm07.yrfcuj (mgr.14201) 809 : audit [DBG] from='client.16136 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:22.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:22 vm09 bash[21220]: audit 2026-03-07T07:09:21.223255+0000 mgr.vm07.yrfcuj (mgr.14201) 809 : audit [DBG] from='client.16136 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:22.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:22 vm09 bash[21220]: audit 2026-03-07T07:09:21.403633+0000 mgr.vm07.yrfcuj (mgr.14201) 810 : audit [DBG] from='client.16140 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:22.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:22 vm09 bash[21220]: audit 2026-03-07T07:09:21.403633+0000 mgr.vm07.yrfcuj (mgr.14201) 810 : audit [DBG] from='client.16140 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:22.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:22 vm09 bash[21220]: cluster 2026-03-07T07:09:21.501117+0000 mgr.vm07.yrfcuj (mgr.14201) 811 : cluster [DBG] pgmap v480: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 200 B/s rd, 401 B/s wr, 0 op/s 2026-03-07T08:09:22.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:22 vm09 bash[21220]: cluster 2026-03-07T07:09:21.501117+0000 mgr.vm07.yrfcuj (mgr.14201) 811 : cluster [DBG] pgmap v480: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 200 B/s rd, 401 B/s wr, 0 op/s 2026-03-07T08:09:22.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:22 vm09 bash[21220]: audit 2026-03-07T07:09:21.605435+0000 mon.vm07 (mon.0) 1105 : audit [DBG] from='client.? 192.168.123.107:0/1467569476' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:22.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:22 vm09 bash[21220]: audit 2026-03-07T07:09:21.605435+0000 mon.vm07 (mon.0) 1105 : audit [DBG] from='client.? 192.168.123.107:0/1467569476' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:22.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:22 vm07 bash[17031]: audit 2026-03-07T07:09:21.223255+0000 mgr.vm07.yrfcuj (mgr.14201) 809 : audit [DBG] from='client.16136 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:22.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:22 vm07 bash[17031]: audit 2026-03-07T07:09:21.223255+0000 mgr.vm07.yrfcuj (mgr.14201) 809 : audit [DBG] from='client.16136 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:22.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:22 vm07 bash[17031]: audit 2026-03-07T07:09:21.403633+0000 mgr.vm07.yrfcuj (mgr.14201) 810 : audit [DBG] from='client.16140 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:22.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:22 vm07 bash[17031]: audit 2026-03-07T07:09:21.403633+0000 mgr.vm07.yrfcuj (mgr.14201) 810 : audit [DBG] from='client.16140 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:22.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:22 vm07 bash[17031]: cluster 2026-03-07T07:09:21.501117+0000 mgr.vm07.yrfcuj (mgr.14201) 811 : cluster [DBG] pgmap v480: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 200 B/s rd, 401 B/s wr, 0 op/s 2026-03-07T08:09:22.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:22 vm07 bash[17031]: cluster 2026-03-07T07:09:21.501117+0000 mgr.vm07.yrfcuj (mgr.14201) 811 : cluster [DBG] pgmap v480: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 200 B/s rd, 401 B/s wr, 0 op/s 2026-03-07T08:09:22.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:22 vm07 bash[17031]: audit 2026-03-07T07:09:21.605435+0000 mon.vm07 (mon.0) 1105 : audit [DBG] from='client.? 192.168.123.107:0/1467569476' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:22.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:22 vm07 bash[17031]: audit 2026-03-07T07:09:21.605435+0000 mon.vm07 (mon.0) 1105 : audit [DBG] from='client.? 192.168.123.107:0/1467569476' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:24.616 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:24 vm09 bash[21220]: cluster 2026-03-07T07:09:23.501639+0000 mgr.vm07.yrfcuj (mgr.14201) 812 : cluster [DBG] pgmap v481: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 100 B/s rd, 200 B/s wr, 0 op/s 2026-03-07T08:09:24.616 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:24 vm09 bash[21220]: cluster 2026-03-07T07:09:23.501639+0000 mgr.vm07.yrfcuj (mgr.14201) 812 : cluster [DBG] pgmap v481: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 100 B/s rd, 200 B/s wr, 0 op/s 2026-03-07T08:09:24.616 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:24 vm09 bash[21220]: audit 2026-03-07T07:09:24.376008+0000 mon.vm07 (mon.0) 1106 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:24.616 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:24 vm09 bash[21220]: audit 2026-03-07T07:09:24.376008+0000 mon.vm07 (mon.0) 1106 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:24.616 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:24 vm09 bash[21220]: audit 2026-03-07T07:09:24.381802+0000 mon.vm07 (mon.0) 1107 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:24.616 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:24 vm09 bash[21220]: audit 2026-03-07T07:09:24.381802+0000 mon.vm07 (mon.0) 1107 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:24.616 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:24 vm09 bash[21220]: audit 2026-03-07T07:09:24.411870+0000 mon.vm07 (mon.0) 1108 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:09:24.616 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:24 vm09 bash[21220]: audit 2026-03-07T07:09:24.411870+0000 mon.vm07 (mon.0) 1108 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:09:24.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:24 vm07 bash[17031]: cluster 2026-03-07T07:09:23.501639+0000 mgr.vm07.yrfcuj (mgr.14201) 812 : cluster [DBG] pgmap v481: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 100 B/s rd, 200 B/s wr, 0 op/s 2026-03-07T08:09:24.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:24 vm07 bash[17031]: cluster 2026-03-07T07:09:23.501639+0000 mgr.vm07.yrfcuj (mgr.14201) 812 : cluster [DBG] pgmap v481: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 100 B/s rd, 200 B/s wr, 0 op/s 2026-03-07T08:09:24.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:24 vm07 bash[17031]: audit 2026-03-07T07:09:24.376008+0000 mon.vm07 (mon.0) 1106 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:24.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:24 vm07 bash[17031]: audit 2026-03-07T07:09:24.376008+0000 mon.vm07 (mon.0) 1106 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:24.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:24 vm07 bash[17031]: audit 2026-03-07T07:09:24.381802+0000 mon.vm07 (mon.0) 1107 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:24.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:24 vm07 bash[17031]: audit 2026-03-07T07:09:24.381802+0000 mon.vm07 (mon.0) 1107 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:24.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:24 vm07 bash[17031]: audit 2026-03-07T07:09:24.411870+0000 mon.vm07 (mon.0) 1108 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:09:24.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:24 vm07 bash[17031]: audit 2026-03-07T07:09:24.411870+0000 mon.vm07 (mon.0) 1108 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:09:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:25 vm09 bash[21220]: audit 2026-03-07T07:09:25.143197+0000 mon.vm07 (mon.0) 1109 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:09:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:25 vm09 bash[21220]: audit 2026-03-07T07:09:25.143197+0000 mon.vm07 (mon.0) 1109 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:09:25.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:25 vm07 bash[17031]: audit 2026-03-07T07:09:25.143197+0000 mon.vm07 (mon.0) 1109 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:09:25.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:25 vm07 bash[17031]: audit 2026-03-07T07:09:25.143197+0000 mon.vm07 (mon.0) 1109 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:09:26.786 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.fgzozy to start 2026-03-07T08:09:26.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:26 vm09 bash[21220]: cluster 2026-03-07T07:09:25.502624+0000 mgr.vm07.yrfcuj (mgr.14201) 813 : cluster [DBG] pgmap v482: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 401 B/s rd, 200 B/s wr, 0 op/s 2026-03-07T08:09:26.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:26 vm09 bash[21220]: cluster 2026-03-07T07:09:25.502624+0000 mgr.vm07.yrfcuj (mgr.14201) 813 : cluster [DBG] pgmap v482: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 401 B/s rd, 200 B/s wr, 0 op/s 2026-03-07T08:09:26.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:26 vm07 bash[17031]: cluster 2026-03-07T07:09:25.502624+0000 mgr.vm07.yrfcuj (mgr.14201) 813 : cluster [DBG] pgmap v482: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 401 B/s rd, 200 B/s wr, 0 op/s 2026-03-07T08:09:26.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:26 vm07 bash[17031]: cluster 2026-03-07T07:09:25.502624+0000 mgr.vm07.yrfcuj (mgr.14201) 813 : cluster [DBG] pgmap v482: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 401 B/s rd, 200 B/s wr, 0 op/s 2026-03-07T08:09:26.942 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:09:26.942 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (5m) 5m ago 11m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:09:26.942 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 5m ago 11m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:09:26.942 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 error 4m ago 11m - - 2026-03-07T08:09:26.942 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (11m) 4m ago 11m 106M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:09:27.140 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:09:27.140 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:09:27.140 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.fgzozy on vm09 is in error state 2026-03-07T08:09:27.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:27 vm09 bash[21220]: audit 2026-03-07T07:09:27.143888+0000 mon.vm07 (mon.0) 1110 : audit [DBG] from='client.? 192.168.123.107:0/2843332287' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:27.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:27 vm09 bash[21220]: audit 2026-03-07T07:09:27.143888+0000 mon.vm07 (mon.0) 1110 : audit [DBG] from='client.? 192.168.123.107:0/2843332287' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:27.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:27 vm07 bash[17031]: audit 2026-03-07T07:09:27.143888+0000 mon.vm07 (mon.0) 1110 : audit [DBG] from='client.? 192.168.123.107:0/2843332287' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:27.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:27 vm07 bash[17031]: audit 2026-03-07T07:09:27.143888+0000 mon.vm07 (mon.0) 1110 : audit [DBG] from='client.? 192.168.123.107:0/2843332287' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:28 vm09 bash[21220]: audit 2026-03-07T07:09:26.773976+0000 mgr.vm07.yrfcuj (mgr.14201) 814 : audit [DBG] from='client.16160 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:28 vm09 bash[21220]: audit 2026-03-07T07:09:26.773976+0000 mgr.vm07.yrfcuj (mgr.14201) 814 : audit [DBG] from='client.16160 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:28 vm09 bash[21220]: audit 2026-03-07T07:09:26.944020+0000 mgr.vm07.yrfcuj (mgr.14201) 815 : audit [DBG] from='client.16164 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:28 vm09 bash[21220]: audit 2026-03-07T07:09:26.944020+0000 mgr.vm07.yrfcuj (mgr.14201) 815 : audit [DBG] from='client.16164 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:28 vm09 bash[21220]: cluster 2026-03-07T07:09:27.503175+0000 mgr.vm07.yrfcuj (mgr.14201) 816 : cluster [DBG] pgmap v483: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 19 KiB/s rd, 170 B/s wr, 29 op/s 2026-03-07T08:09:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:28 vm09 bash[21220]: cluster 2026-03-07T07:09:27.503175+0000 mgr.vm07.yrfcuj (mgr.14201) 816 : cluster [DBG] pgmap v483: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 19 KiB/s rd, 170 B/s wr, 29 op/s 2026-03-07T08:09:28.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:28 vm07 bash[17031]: audit 2026-03-07T07:09:26.773976+0000 mgr.vm07.yrfcuj (mgr.14201) 814 : audit [DBG] from='client.16160 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:28.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:28 vm07 bash[17031]: audit 2026-03-07T07:09:26.773976+0000 mgr.vm07.yrfcuj (mgr.14201) 814 : audit [DBG] from='client.16160 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:28.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:28 vm07 bash[17031]: audit 2026-03-07T07:09:26.944020+0000 mgr.vm07.yrfcuj (mgr.14201) 815 : audit [DBG] from='client.16164 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:28.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:28 vm07 bash[17031]: audit 2026-03-07T07:09:26.944020+0000 mgr.vm07.yrfcuj (mgr.14201) 815 : audit [DBG] from='client.16164 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:28.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:28 vm07 bash[17031]: cluster 2026-03-07T07:09:27.503175+0000 mgr.vm07.yrfcuj (mgr.14201) 816 : cluster [DBG] pgmap v483: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 19 KiB/s rd, 170 B/s wr, 29 op/s 2026-03-07T08:09:28.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:28 vm07 bash[17031]: cluster 2026-03-07T07:09:27.503175+0000 mgr.vm07.yrfcuj (mgr.14201) 816 : cluster [DBG] pgmap v483: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 19 KiB/s rd, 170 B/s wr, 29 op/s 2026-03-07T08:09:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:29 vm09 bash[21220]: audit 2026-03-07T07:09:28.809684+0000 mon.vm07 (mon.0) 1111 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:29 vm09 bash[21220]: audit 2026-03-07T07:09:28.809684+0000 mon.vm07 (mon.0) 1111 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:29 vm09 bash[21220]: audit 2026-03-07T07:09:28.813980+0000 mon.vm07 (mon.0) 1112 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:29 vm09 bash[21220]: audit 2026-03-07T07:09:28.813980+0000 mon.vm07 (mon.0) 1112 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:29 vm09 bash[21220]: audit 2026-03-07T07:09:28.814814+0000 mon.vm07 (mon.0) 1113 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:09:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:29 vm09 bash[21220]: audit 2026-03-07T07:09:28.814814+0000 mon.vm07 (mon.0) 1113 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:09:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:29 vm09 bash[21220]: audit 2026-03-07T07:09:28.815751+0000 mon.vm07 (mon.0) 1114 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:09:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:29 vm09 bash[21220]: audit 2026-03-07T07:09:28.815751+0000 mon.vm07 (mon.0) 1114 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:09:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:29 vm09 bash[21220]: cluster 2026-03-07T07:09:28.817079+0000 mgr.vm07.yrfcuj (mgr.14201) 817 : cluster [DBG] pgmap v484: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 49 KiB/s rd, 0 B/s wr, 75 op/s 2026-03-07T08:09:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:29 vm09 bash[21220]: cluster 2026-03-07T07:09:28.817079+0000 mgr.vm07.yrfcuj (mgr.14201) 817 : cluster [DBG] pgmap v484: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 49 KiB/s rd, 0 B/s wr, 75 op/s 2026-03-07T08:09:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:29 vm09 bash[21220]: audit 2026-03-07T07:09:28.819812+0000 mon.vm07 (mon.0) 1115 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:29 vm09 bash[21220]: audit 2026-03-07T07:09:28.819812+0000 mon.vm07 (mon.0) 1115 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:29 vm09 bash[21220]: audit 2026-03-07T07:09:28.821364+0000 mon.vm07 (mon.0) 1116 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:09:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:29 vm09 bash[21220]: audit 2026-03-07T07:09:28.821364+0000 mon.vm07 (mon.0) 1116 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:09:30.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:29 vm07 bash[17031]: audit 2026-03-07T07:09:28.809684+0000 mon.vm07 (mon.0) 1111 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:30.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:29 vm07 bash[17031]: audit 2026-03-07T07:09:28.809684+0000 mon.vm07 (mon.0) 1111 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:30.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:29 vm07 bash[17031]: audit 2026-03-07T07:09:28.813980+0000 mon.vm07 (mon.0) 1112 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:30.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:29 vm07 bash[17031]: audit 2026-03-07T07:09:28.813980+0000 mon.vm07 (mon.0) 1112 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:30.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:29 vm07 bash[17031]: audit 2026-03-07T07:09:28.814814+0000 mon.vm07 (mon.0) 1113 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:09:30.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:29 vm07 bash[17031]: audit 2026-03-07T07:09:28.814814+0000 mon.vm07 (mon.0) 1113 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:09:30.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:29 vm07 bash[17031]: audit 2026-03-07T07:09:28.815751+0000 mon.vm07 (mon.0) 1114 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:09:30.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:29 vm07 bash[17031]: audit 2026-03-07T07:09:28.815751+0000 mon.vm07 (mon.0) 1114 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:09:30.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:29 vm07 bash[17031]: cluster 2026-03-07T07:09:28.817079+0000 mgr.vm07.yrfcuj (mgr.14201) 817 : cluster [DBG] pgmap v484: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 49 KiB/s rd, 0 B/s wr, 75 op/s 2026-03-07T08:09:30.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:29 vm07 bash[17031]: cluster 2026-03-07T07:09:28.817079+0000 mgr.vm07.yrfcuj (mgr.14201) 817 : cluster [DBG] pgmap v484: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 49 KiB/s rd, 0 B/s wr, 75 op/s 2026-03-07T08:09:30.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:29 vm07 bash[17031]: audit 2026-03-07T07:09:28.819812+0000 mon.vm07 (mon.0) 1115 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:30.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:29 vm07 bash[17031]: audit 2026-03-07T07:09:28.819812+0000 mon.vm07 (mon.0) 1115 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:30.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:29 vm07 bash[17031]: audit 2026-03-07T07:09:28.821364+0000 mon.vm07 (mon.0) 1116 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:09:30.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:29 vm07 bash[17031]: audit 2026-03-07T07:09:28.821364+0000 mon.vm07 (mon.0) 1116 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:09:31.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:30 vm09 bash[21220]: cluster 2026-03-07T07:09:29.814667+0000 mon.vm07 (mon.0) 1117 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-07T08:09:31.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:30 vm09 bash[21220]: cluster 2026-03-07T07:09:29.814667+0000 mon.vm07 (mon.0) 1117 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-07T08:09:31.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:30 vm09 bash[21220]: cluster 2026-03-07T07:09:29.814676+0000 mon.vm07 (mon.0) 1118 : cluster [INF] Cluster is now healthy 2026-03-07T08:09:31.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:30 vm09 bash[21220]: cluster 2026-03-07T07:09:29.814676+0000 mon.vm07 (mon.0) 1118 : cluster [INF] Cluster is now healthy 2026-03-07T08:09:31.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:30 vm07 bash[17031]: cluster 2026-03-07T07:09:29.814667+0000 mon.vm07 (mon.0) 1117 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-07T08:09:31.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:30 vm07 bash[17031]: cluster 2026-03-07T07:09:29.814667+0000 mon.vm07 (mon.0) 1117 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-07T08:09:31.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:30 vm07 bash[17031]: cluster 2026-03-07T07:09:29.814676+0000 mon.vm07 (mon.0) 1118 : cluster [INF] Cluster is now healthy 2026-03-07T08:09:31.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:30 vm07 bash[17031]: cluster 2026-03-07T07:09:29.814676+0000 mon.vm07 (mon.0) 1118 : cluster [INF] Cluster is now healthy 2026-03-07T08:09:32.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:31 vm09 bash[21220]: cluster 2026-03-07T07:09:30.817580+0000 mgr.vm07.yrfcuj (mgr.14201) 818 : cluster [DBG] pgmap v485: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 0 B/s wr, 87 op/s 2026-03-07T08:09:32.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:31 vm09 bash[21220]: cluster 2026-03-07T07:09:30.817580+0000 mgr.vm07.yrfcuj (mgr.14201) 818 : cluster [DBG] pgmap v485: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 0 B/s wr, 87 op/s 2026-03-07T08:09:32.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:31 vm07 bash[17031]: cluster 2026-03-07T07:09:30.817580+0000 mgr.vm07.yrfcuj (mgr.14201) 818 : cluster [DBG] pgmap v485: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 0 B/s wr, 87 op/s 2026-03-07T08:09:32.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:31 vm07 bash[17031]: cluster 2026-03-07T07:09:30.817580+0000 mgr.vm07.yrfcuj (mgr.14201) 818 : cluster [DBG] pgmap v485: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 0 B/s wr, 87 op/s 2026-03-07T08:09:32.332 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (7s) 3s ago 11m 89.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:09:32.502 INFO:teuthology.orchestra.run.vm07.stdout:Scheduled to stop rgw.foo.vm09.kpgoql on host 'vm09' 2026-03-07T08:09:32.700 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:09:32.871 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:09:32.871 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (6m) 5m ago 11m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:09:32.871 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 5m ago 11m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:09:32.871 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (8s) 4s ago 11m 89.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:09:32.871 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (11m) 4s ago 11m 108M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:09:33.082 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:09:33.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:32.321445+0000 mgr.vm07.yrfcuj (mgr.14201) 819 : audit [DBG] from='client.16172 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:33.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:32.321445+0000 mgr.vm07.yrfcuj (mgr.14201) 819 : audit [DBG] from='client.16172 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:33.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:32.491695+0000 mgr.vm07.yrfcuj (mgr.14201) 820 : audit [DBG] from='client.16176 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "rgw.foo.vm09.kpgoql", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:33.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:32.491695+0000 mgr.vm07.yrfcuj (mgr.14201) 820 : audit [DBG] from='client.16176 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "rgw.foo.vm09.kpgoql", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:33.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: cephadm 2026-03-07T07:09:32.492077+0000 mgr.vm07.yrfcuj (mgr.14201) 821 : cephadm [INF] Schedule stop daemon rgw.foo.vm09.kpgoql 2026-03-07T08:09:33.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: cephadm 2026-03-07T07:09:32.492077+0000 mgr.vm07.yrfcuj (mgr.14201) 821 : cephadm [INF] Schedule stop daemon rgw.foo.vm09.kpgoql 2026-03-07T08:09:33.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:32.498747+0000 mon.vm07 (mon.0) 1119 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:33.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:32.498747+0000 mon.vm07 (mon.0) 1119 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:33.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:32.504101+0000 mon.vm07 (mon.0) 1120 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:33.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:32.504101+0000 mon.vm07 (mon.0) 1120 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:33.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:32.505305+0000 mon.vm07 (mon.0) 1121 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:09:33.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:32.505305+0000 mon.vm07 (mon.0) 1121 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:09:33.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:32.507307+0000 mon.vm07 (mon.0) 1122 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:09:33.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:32.507307+0000 mon.vm07 (mon.0) 1122 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:09:33.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:32.508214+0000 mon.vm07 (mon.0) 1123 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:09:33.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:32.508214+0000 mon.vm07 (mon.0) 1123 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:09:33.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:32.513038+0000 mon.vm07 (mon.0) 1124 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:33.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:32.513038+0000 mon.vm07 (mon.0) 1124 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:33.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:32.514716+0000 mon.vm07 (mon.0) 1125 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:09:33.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:32.514716+0000 mon.vm07 (mon.0) 1125 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:09:33.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:33.086464+0000 mon.vm07 (mon.0) 1126 : audit [DBG] from='client.? 192.168.123.107:0/3651931866' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:33.866 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:33 vm09 bash[21220]: audit 2026-03-07T07:09:33.086464+0000 mon.vm07 (mon.0) 1126 : audit [DBG] from='client.? 192.168.123.107:0/3651931866' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:32.321445+0000 mgr.vm07.yrfcuj (mgr.14201) 819 : audit [DBG] from='client.16172 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:32.321445+0000 mgr.vm07.yrfcuj (mgr.14201) 819 : audit [DBG] from='client.16172 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:32.491695+0000 mgr.vm07.yrfcuj (mgr.14201) 820 : audit [DBG] from='client.16176 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "rgw.foo.vm09.kpgoql", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:32.491695+0000 mgr.vm07.yrfcuj (mgr.14201) 820 : audit [DBG] from='client.16176 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "rgw.foo.vm09.kpgoql", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: cephadm 2026-03-07T07:09:32.492077+0000 mgr.vm07.yrfcuj (mgr.14201) 821 : cephadm [INF] Schedule stop daemon rgw.foo.vm09.kpgoql 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: cephadm 2026-03-07T07:09:32.492077+0000 mgr.vm07.yrfcuj (mgr.14201) 821 : cephadm [INF] Schedule stop daemon rgw.foo.vm09.kpgoql 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:32.498747+0000 mon.vm07 (mon.0) 1119 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:32.498747+0000 mon.vm07 (mon.0) 1119 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:32.504101+0000 mon.vm07 (mon.0) 1120 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:32.504101+0000 mon.vm07 (mon.0) 1120 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:32.505305+0000 mon.vm07 (mon.0) 1121 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:32.505305+0000 mon.vm07 (mon.0) 1121 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:32.507307+0000 mon.vm07 (mon.0) 1122 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:32.507307+0000 mon.vm07 (mon.0) 1122 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:32.508214+0000 mon.vm07 (mon.0) 1123 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:32.508214+0000 mon.vm07 (mon.0) 1123 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:32.513038+0000 mon.vm07 (mon.0) 1124 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:32.513038+0000 mon.vm07 (mon.0) 1124 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:32.514716+0000 mon.vm07 (mon.0) 1125 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:32.514716+0000 mon.vm07 (mon.0) 1125 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:33.086464+0000 mon.vm07 (mon.0) 1126 : audit [DBG] from='client.? 192.168.123.107:0/3651931866' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:33.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:33 vm07 bash[17031]: audit 2026-03-07T07:09:33.086464+0000 mon.vm07 (mon.0) 1126 : audit [DBG] from='client.? 192.168.123.107:0/3651931866' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:34.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:34 vm09 bash[21220]: audit 2026-03-07T07:09:32.688394+0000 mgr.vm07.yrfcuj (mgr.14201) 822 : audit [DBG] from='client.16180 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:34.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:34 vm09 bash[21220]: audit 2026-03-07T07:09:32.688394+0000 mgr.vm07.yrfcuj (mgr.14201) 822 : audit [DBG] from='client.16180 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:34.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:34 vm09 bash[21220]: cluster 2026-03-07T07:09:32.818028+0000 mgr.vm07.yrfcuj (mgr.14201) 823 : cluster [DBG] pgmap v486: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 180 B/s wr, 87 op/s 2026-03-07T08:09:34.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:34 vm09 bash[21220]: cluster 2026-03-07T07:09:32.818028+0000 mgr.vm07.yrfcuj (mgr.14201) 823 : cluster [DBG] pgmap v486: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 180 B/s wr, 87 op/s 2026-03-07T08:09:34.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:34 vm09 bash[21220]: audit 2026-03-07T07:09:32.871838+0000 mgr.vm07.yrfcuj (mgr.14201) 824 : audit [DBG] from='client.16184 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:34.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:34 vm09 bash[21220]: audit 2026-03-07T07:09:32.871838+0000 mgr.vm07.yrfcuj (mgr.14201) 824 : audit [DBG] from='client.16184 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:34.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:34 vm07 bash[17031]: audit 2026-03-07T07:09:32.688394+0000 mgr.vm07.yrfcuj (mgr.14201) 822 : audit [DBG] from='client.16180 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:34.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:34 vm07 bash[17031]: audit 2026-03-07T07:09:32.688394+0000 mgr.vm07.yrfcuj (mgr.14201) 822 : audit [DBG] from='client.16180 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:34.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:34 vm07 bash[17031]: cluster 2026-03-07T07:09:32.818028+0000 mgr.vm07.yrfcuj (mgr.14201) 823 : cluster [DBG] pgmap v486: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 180 B/s wr, 87 op/s 2026-03-07T08:09:34.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:34 vm07 bash[17031]: cluster 2026-03-07T07:09:32.818028+0000 mgr.vm07.yrfcuj (mgr.14201) 823 : cluster [DBG] pgmap v486: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 180 B/s wr, 87 op/s 2026-03-07T08:09:34.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:34 vm07 bash[17031]: audit 2026-03-07T07:09:32.871838+0000 mgr.vm07.yrfcuj (mgr.14201) 824 : audit [DBG] from='client.16184 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:34.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:34 vm07 bash[17031]: audit 2026-03-07T07:09:32.871838+0000 mgr.vm07.yrfcuj (mgr.14201) 824 : audit [DBG] from='client.16184 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:36 vm09 bash[21220]: cluster 2026-03-07T07:09:34.818517+0000 mgr.vm07.yrfcuj (mgr.14201) 825 : cluster [DBG] pgmap v487: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 180 B/s wr, 87 op/s 2026-03-07T08:09:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:36 vm09 bash[21220]: cluster 2026-03-07T07:09:34.818517+0000 mgr.vm07.yrfcuj (mgr.14201) 825 : cluster [DBG] pgmap v487: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 180 B/s wr, 87 op/s 2026-03-07T08:09:36.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:36 vm07 bash[17031]: cluster 2026-03-07T07:09:34.818517+0000 mgr.vm07.yrfcuj (mgr.14201) 825 : cluster [DBG] pgmap v487: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 180 B/s wr, 87 op/s 2026-03-07T08:09:36.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:36 vm07 bash[17031]: cluster 2026-03-07T07:09:34.818517+0000 mgr.vm07.yrfcuj (mgr.14201) 825 : cluster [DBG] pgmap v487: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 180 B/s wr, 87 op/s 2026-03-07T08:09:38.272 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:09:38.434 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:09:38.435 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (6m) 5m ago 12m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:09:38.435 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 5m ago 11m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:09:38.435 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (14s) 9s ago 12m 89.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:09:38.435 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (11m) 9s ago 11m 108M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:09:38.639 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:09:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:38 vm09 bash[21220]: cluster 2026-03-07T07:09:36.818933+0000 mgr.vm07.yrfcuj (mgr.14201) 826 : cluster [DBG] pgmap v488: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 180 B/s wr, 87 op/s 2026-03-07T08:09:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:38 vm09 bash[21220]: cluster 2026-03-07T07:09:36.818933+0000 mgr.vm07.yrfcuj (mgr.14201) 826 : cluster [DBG] pgmap v488: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 180 B/s wr, 87 op/s 2026-03-07T08:09:38.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:38 vm07 bash[17031]: cluster 2026-03-07T07:09:36.818933+0000 mgr.vm07.yrfcuj (mgr.14201) 826 : cluster [DBG] pgmap v488: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 180 B/s wr, 87 op/s 2026-03-07T08:09:38.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:38 vm07 bash[17031]: cluster 2026-03-07T07:09:36.818933+0000 mgr.vm07.yrfcuj (mgr.14201) 826 : cluster [DBG] pgmap v488: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 180 B/s wr, 87 op/s 2026-03-07T08:09:39.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:39 vm09 bash[21220]: audit 2026-03-07T07:09:38.259548+0000 mgr.vm07.yrfcuj (mgr.14201) 827 : audit [DBG] from='client.16192 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:39.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:39 vm09 bash[21220]: audit 2026-03-07T07:09:38.259548+0000 mgr.vm07.yrfcuj (mgr.14201) 827 : audit [DBG] from='client.16192 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:39.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:39 vm09 bash[21220]: audit 2026-03-07T07:09:38.436191+0000 mgr.vm07.yrfcuj (mgr.14201) 828 : audit [DBG] from='client.16196 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:39.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:39 vm09 bash[21220]: audit 2026-03-07T07:09:38.436191+0000 mgr.vm07.yrfcuj (mgr.14201) 828 : audit [DBG] from='client.16196 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:39.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:39 vm09 bash[21220]: audit 2026-03-07T07:09:38.643124+0000 mon.vm07 (mon.0) 1127 : audit [DBG] from='client.? 192.168.123.107:0/2951975362' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:39.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:39 vm09 bash[21220]: audit 2026-03-07T07:09:38.643124+0000 mon.vm07 (mon.0) 1127 : audit [DBG] from='client.? 192.168.123.107:0/2951975362' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:39.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:39 vm09 bash[21220]: cluster 2026-03-07T07:09:38.819489+0000 mgr.vm07.yrfcuj (mgr.14201) 829 : cluster [DBG] pgmap v489: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 180 B/s wr, 56 op/s 2026-03-07T08:09:39.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:39 vm09 bash[21220]: cluster 2026-03-07T07:09:38.819489+0000 mgr.vm07.yrfcuj (mgr.14201) 829 : cluster [DBG] pgmap v489: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 180 B/s wr, 56 op/s 2026-03-07T08:09:39.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:39 vm07 bash[17031]: audit 2026-03-07T07:09:38.259548+0000 mgr.vm07.yrfcuj (mgr.14201) 827 : audit [DBG] from='client.16192 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:39.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:39 vm07 bash[17031]: audit 2026-03-07T07:09:38.259548+0000 mgr.vm07.yrfcuj (mgr.14201) 827 : audit [DBG] from='client.16192 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:39.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:39 vm07 bash[17031]: audit 2026-03-07T07:09:38.436191+0000 mgr.vm07.yrfcuj (mgr.14201) 828 : audit [DBG] from='client.16196 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:39.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:39 vm07 bash[17031]: audit 2026-03-07T07:09:38.436191+0000 mgr.vm07.yrfcuj (mgr.14201) 828 : audit [DBG] from='client.16196 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:39.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:39 vm07 bash[17031]: audit 2026-03-07T07:09:38.643124+0000 mon.vm07 (mon.0) 1127 : audit [DBG] from='client.? 192.168.123.107:0/2951975362' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:39.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:39 vm07 bash[17031]: audit 2026-03-07T07:09:38.643124+0000 mon.vm07 (mon.0) 1127 : audit [DBG] from='client.? 192.168.123.107:0/2951975362' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:39.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:39 vm07 bash[17031]: cluster 2026-03-07T07:09:38.819489+0000 mgr.vm07.yrfcuj (mgr.14201) 829 : cluster [DBG] pgmap v489: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 180 B/s wr, 56 op/s 2026-03-07T08:09:39.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:39 vm07 bash[17031]: cluster 2026-03-07T07:09:38.819489+0000 mgr.vm07.yrfcuj (mgr.14201) 829 : cluster [DBG] pgmap v489: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 180 B/s wr, 56 op/s 2026-03-07T08:09:41.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:41 vm07 bash[17031]: audit 2026-03-07T07:09:40.145703+0000 mon.vm07 (mon.0) 1128 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:41.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:41 vm07 bash[17031]: audit 2026-03-07T07:09:40.145703+0000 mon.vm07 (mon.0) 1128 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:41.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:41 vm07 bash[17031]: audit 2026-03-07T07:09:40.146446+0000 mon.vm07 (mon.0) 1129 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:09:41.399 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:41 vm07 bash[17031]: audit 2026-03-07T07:09:40.146446+0000 mon.vm07 (mon.0) 1129 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:09:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:41 vm09 bash[21220]: audit 2026-03-07T07:09:40.145703+0000 mon.vm07 (mon.0) 1128 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:41 vm09 bash[21220]: audit 2026-03-07T07:09:40.145703+0000 mon.vm07 (mon.0) 1128 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:41 vm09 bash[21220]: audit 2026-03-07T07:09:40.146446+0000 mon.vm07 (mon.0) 1129 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:09:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:41 vm09 bash[21220]: audit 2026-03-07T07:09:40.146446+0000 mon.vm07 (mon.0) 1129 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:09:42.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:42 vm09 bash[21220]: cluster 2026-03-07T07:09:40.820014+0000 mgr.vm07.yrfcuj (mgr.14201) 830 : cluster [DBG] pgmap v490: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 6.8 KiB/s rd, 170 B/s wr, 10 op/s 2026-03-07T08:09:42.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:42 vm09 bash[21220]: cluster 2026-03-07T07:09:40.820014+0000 mgr.vm07.yrfcuj (mgr.14201) 830 : cluster [DBG] pgmap v490: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 6.8 KiB/s rd, 170 B/s wr, 10 op/s 2026-03-07T08:09:42.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:42 vm07 bash[17031]: cluster 2026-03-07T07:09:40.820014+0000 mgr.vm07.yrfcuj (mgr.14201) 830 : cluster [DBG] pgmap v490: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 6.8 KiB/s rd, 170 B/s wr, 10 op/s 2026-03-07T08:09:42.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:42 vm07 bash[17031]: cluster 2026-03-07T07:09:40.820014+0000 mgr.vm07.yrfcuj (mgr.14201) 830 : cluster [DBG] pgmap v490: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 6.8 KiB/s rd, 170 B/s wr, 10 op/s 2026-03-07T08:09:43.829 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:09:43.984 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:09:43.984 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (6m) 5m ago 12m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:09:43.984 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 5m ago 12m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:09:43.984 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (19s) 15s ago 12m 89.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:09:43.984 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (12m) 15s ago 12m 108M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:09:44.181 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:09:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:44 vm09 bash[21220]: cluster 2026-03-07T07:09:42.820570+0000 mgr.vm07.yrfcuj (mgr.14201) 831 : cluster [DBG] pgmap v491: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:09:44.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:44 vm09 bash[21220]: cluster 2026-03-07T07:09:42.820570+0000 mgr.vm07.yrfcuj (mgr.14201) 831 : cluster [DBG] pgmap v491: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:09:44.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:44 vm07 bash[17031]: cluster 2026-03-07T07:09:42.820570+0000 mgr.vm07.yrfcuj (mgr.14201) 831 : cluster [DBG] pgmap v491: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:09:44.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:44 vm07 bash[17031]: cluster 2026-03-07T07:09:42.820570+0000 mgr.vm07.yrfcuj (mgr.14201) 831 : cluster [DBG] pgmap v491: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:09:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:45 vm09 bash[21220]: audit 2026-03-07T07:09:43.814820+0000 mgr.vm07.yrfcuj (mgr.14201) 832 : audit [DBG] from='client.16204 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:45 vm09 bash[21220]: audit 2026-03-07T07:09:43.814820+0000 mgr.vm07.yrfcuj (mgr.14201) 832 : audit [DBG] from='client.16204 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:45 vm09 bash[21220]: audit 2026-03-07T07:09:43.985230+0000 mgr.vm07.yrfcuj (mgr.14201) 833 : audit [DBG] from='client.16208 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:45 vm09 bash[21220]: audit 2026-03-07T07:09:43.985230+0000 mgr.vm07.yrfcuj (mgr.14201) 833 : audit [DBG] from='client.16208 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:45 vm09 bash[21220]: audit 2026-03-07T07:09:44.184677+0000 mon.vm07 (mon.0) 1130 : audit [DBG] from='client.? 192.168.123.107:0/3914936834' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:45 vm09 bash[21220]: audit 2026-03-07T07:09:44.184677+0000 mon.vm07 (mon.0) 1130 : audit [DBG] from='client.? 192.168.123.107:0/3914936834' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:45.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:45 vm07 bash[17031]: audit 2026-03-07T07:09:43.814820+0000 mgr.vm07.yrfcuj (mgr.14201) 832 : audit [DBG] from='client.16204 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:45.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:45 vm07 bash[17031]: audit 2026-03-07T07:09:43.814820+0000 mgr.vm07.yrfcuj (mgr.14201) 832 : audit [DBG] from='client.16204 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:45.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:45 vm07 bash[17031]: audit 2026-03-07T07:09:43.985230+0000 mgr.vm07.yrfcuj (mgr.14201) 833 : audit [DBG] from='client.16208 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:45.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:45 vm07 bash[17031]: audit 2026-03-07T07:09:43.985230+0000 mgr.vm07.yrfcuj (mgr.14201) 833 : audit [DBG] from='client.16208 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:45.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:45 vm07 bash[17031]: audit 2026-03-07T07:09:44.184677+0000 mon.vm07 (mon.0) 1130 : audit [DBG] from='client.? 192.168.123.107:0/3914936834' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:45.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:45 vm07 bash[17031]: audit 2026-03-07T07:09:44.184677+0000 mon.vm07 (mon.0) 1130 : audit [DBG] from='client.? 192.168.123.107:0/3914936834' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:46.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:46 vm09 bash[21220]: cluster 2026-03-07T07:09:44.820959+0000 mgr.vm07.yrfcuj (mgr.14201) 834 : cluster [DBG] pgmap v492: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:46.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:46 vm09 bash[21220]: cluster 2026-03-07T07:09:44.820959+0000 mgr.vm07.yrfcuj (mgr.14201) 834 : cluster [DBG] pgmap v492: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:46.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:46 vm07 bash[17031]: cluster 2026-03-07T07:09:44.820959+0000 mgr.vm07.yrfcuj (mgr.14201) 834 : cluster [DBG] pgmap v492: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:46.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:46 vm07 bash[17031]: cluster 2026-03-07T07:09:44.820959+0000 mgr.vm07.yrfcuj (mgr.14201) 834 : cluster [DBG] pgmap v492: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:48 vm09 bash[21220]: cluster 2026-03-07T07:09:46.821412+0000 mgr.vm07.yrfcuj (mgr.14201) 835 : cluster [DBG] pgmap v493: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:48 vm09 bash[21220]: cluster 2026-03-07T07:09:46.821412+0000 mgr.vm07.yrfcuj (mgr.14201) 835 : cluster [DBG] pgmap v493: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:48 vm09 bash[21220]: audit 2026-03-07T07:09:47.515307+0000 mon.vm07 (mon.0) 1131 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:48 vm09 bash[21220]: audit 2026-03-07T07:09:47.515307+0000 mon.vm07 (mon.0) 1131 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:48 vm09 bash[21220]: audit 2026-03-07T07:09:47.520572+0000 mon.vm07 (mon.0) 1132 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:48 vm09 bash[21220]: audit 2026-03-07T07:09:47.520572+0000 mon.vm07 (mon.0) 1132 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:48 vm09 bash[21220]: audit 2026-03-07T07:09:47.550078+0000 mon.vm07 (mon.0) 1133 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:09:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:48 vm09 bash[21220]: audit 2026-03-07T07:09:47.550078+0000 mon.vm07 (mon.0) 1133 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:09:48.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:48 vm07 bash[17031]: cluster 2026-03-07T07:09:46.821412+0000 mgr.vm07.yrfcuj (mgr.14201) 835 : cluster [DBG] pgmap v493: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:48.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:48 vm07 bash[17031]: cluster 2026-03-07T07:09:46.821412+0000 mgr.vm07.yrfcuj (mgr.14201) 835 : cluster [DBG] pgmap v493: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:48.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:48 vm07 bash[17031]: audit 2026-03-07T07:09:47.515307+0000 mon.vm07 (mon.0) 1131 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:48.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:48 vm07 bash[17031]: audit 2026-03-07T07:09:47.515307+0000 mon.vm07 (mon.0) 1131 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:48.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:48 vm07 bash[17031]: audit 2026-03-07T07:09:47.520572+0000 mon.vm07 (mon.0) 1132 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:48.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:48 vm07 bash[17031]: audit 2026-03-07T07:09:47.520572+0000 mon.vm07 (mon.0) 1132 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:48.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:48 vm07 bash[17031]: audit 2026-03-07T07:09:47.550078+0000 mon.vm07 (mon.0) 1133 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:09:48.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:48 vm07 bash[17031]: audit 2026-03-07T07:09:47.550078+0000 mon.vm07 (mon.0) 1133 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:09:49.372 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:09:49.533 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:09:49.534 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (6m) 5m ago 12m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:09:49.534 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 5m ago 12m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:09:49.534 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (25s) 20s ago 12m 89.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:09:49.534 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (12m) 20s ago 12m 108M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 53d4fdc6f05b 2026-03-07T08:09:49.729 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:09:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:50 vm09 bash[21220]: cluster 2026-03-07T07:09:48.821842+0000 mgr.vm07.yrfcuj (mgr.14201) 836 : cluster [DBG] pgmap v494: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:50 vm09 bash[21220]: cluster 2026-03-07T07:09:48.821842+0000 mgr.vm07.yrfcuj (mgr.14201) 836 : cluster [DBG] pgmap v494: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:50 vm09 bash[21220]: audit 2026-03-07T07:09:49.361159+0000 mgr.vm07.yrfcuj (mgr.14201) 837 : audit [DBG] from='client.16216 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:50 vm09 bash[21220]: audit 2026-03-07T07:09:49.361159+0000 mgr.vm07.yrfcuj (mgr.14201) 837 : audit [DBG] from='client.16216 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:50 vm09 bash[21220]: audit 2026-03-07T07:09:49.534994+0000 mgr.vm07.yrfcuj (mgr.14201) 838 : audit [DBG] from='client.16220 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:50 vm09 bash[21220]: audit 2026-03-07T07:09:49.534994+0000 mgr.vm07.yrfcuj (mgr.14201) 838 : audit [DBG] from='client.16220 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:50 vm09 bash[21220]: audit 2026-03-07T07:09:49.733363+0000 mon.vm07 (mon.0) 1134 : audit [DBG] from='client.? 192.168.123.107:0/3536394391' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:50 vm09 bash[21220]: audit 2026-03-07T07:09:49.733363+0000 mon.vm07 (mon.0) 1134 : audit [DBG] from='client.? 192.168.123.107:0/3536394391' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:50.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:50 vm07 bash[17031]: cluster 2026-03-07T07:09:48.821842+0000 mgr.vm07.yrfcuj (mgr.14201) 836 : cluster [DBG] pgmap v494: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:50.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:50 vm07 bash[17031]: cluster 2026-03-07T07:09:48.821842+0000 mgr.vm07.yrfcuj (mgr.14201) 836 : cluster [DBG] pgmap v494: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:50.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:50 vm07 bash[17031]: audit 2026-03-07T07:09:49.361159+0000 mgr.vm07.yrfcuj (mgr.14201) 837 : audit [DBG] from='client.16216 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:50.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:50 vm07 bash[17031]: audit 2026-03-07T07:09:49.361159+0000 mgr.vm07.yrfcuj (mgr.14201) 837 : audit [DBG] from='client.16216 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:50.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:50 vm07 bash[17031]: audit 2026-03-07T07:09:49.534994+0000 mgr.vm07.yrfcuj (mgr.14201) 838 : audit [DBG] from='client.16220 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:50.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:50 vm07 bash[17031]: audit 2026-03-07T07:09:49.534994+0000 mgr.vm07.yrfcuj (mgr.14201) 838 : audit [DBG] from='client.16220 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:50.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:50 vm07 bash[17031]: audit 2026-03-07T07:09:49.733363+0000 mon.vm07 (mon.0) 1134 : audit [DBG] from='client.? 192.168.123.107:0/3536394391' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:50.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:50 vm07 bash[17031]: audit 2026-03-07T07:09:49.733363+0000 mon.vm07 (mon.0) 1134 : audit [DBG] from='client.? 192.168.123.107:0/3536394391' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:52 vm09 bash[21220]: cluster 2026-03-07T07:09:50.822326+0000 mgr.vm07.yrfcuj (mgr.14201) 839 : cluster [DBG] pgmap v495: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:52 vm09 bash[21220]: cluster 2026-03-07T07:09:50.822326+0000 mgr.vm07.yrfcuj (mgr.14201) 839 : cluster [DBG] pgmap v495: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:52.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:52 vm07 bash[17031]: cluster 2026-03-07T07:09:50.822326+0000 mgr.vm07.yrfcuj (mgr.14201) 839 : cluster [DBG] pgmap v495: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:52.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:52 vm07 bash[17031]: cluster 2026-03-07T07:09:50.822326+0000 mgr.vm07.yrfcuj (mgr.14201) 839 : cluster [DBG] pgmap v495: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:53 vm09 bash[21220]: cluster 2026-03-07T07:09:52.823665+0000 mgr.vm07.yrfcuj (mgr.14201) 840 : cluster [DBG] pgmap v496: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:53 vm09 bash[21220]: cluster 2026-03-07T07:09:52.823665+0000 mgr.vm07.yrfcuj (mgr.14201) 840 : cluster [DBG] pgmap v496: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:53 vm09 bash[21220]: audit 2026-03-07T07:09:52.827312+0000 mon.vm07 (mon.0) 1135 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:53 vm09 bash[21220]: audit 2026-03-07T07:09:52.827312+0000 mon.vm07 (mon.0) 1135 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:53 vm09 bash[21220]: audit 2026-03-07T07:09:52.831554+0000 mon.vm07 (mon.0) 1136 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:53 vm09 bash[21220]: audit 2026-03-07T07:09:52.831554+0000 mon.vm07 (mon.0) 1136 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:53 vm09 bash[21220]: audit 2026-03-07T07:09:53.150399+0000 mon.vm07 (mon.0) 1137 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:09:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:53 vm09 bash[21220]: audit 2026-03-07T07:09:53.150399+0000 mon.vm07 (mon.0) 1137 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:09:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:53 vm09 bash[21220]: audit 2026-03-07T07:09:53.150995+0000 mon.vm07 (mon.0) 1138 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:09:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:53 vm09 bash[21220]: audit 2026-03-07T07:09:53.150995+0000 mon.vm07 (mon.0) 1138 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:09:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:53 vm09 bash[21220]: cluster 2026-03-07T07:09:53.151923+0000 mgr.vm07.yrfcuj (mgr.14201) 841 : cluster [DBG] pgmap v497: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:09:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:53 vm09 bash[21220]: cluster 2026-03-07T07:09:53.151923+0000 mgr.vm07.yrfcuj (mgr.14201) 841 : cluster [DBG] pgmap v497: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:09:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:53 vm09 bash[21220]: audit 2026-03-07T07:09:53.156385+0000 mon.vm07 (mon.0) 1139 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:53 vm09 bash[21220]: audit 2026-03-07T07:09:53.156385+0000 mon.vm07 (mon.0) 1139 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:53 vm09 bash[21220]: audit 2026-03-07T07:09:53.157912+0000 mon.vm07 (mon.0) 1140 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:09:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:53 vm09 bash[21220]: audit 2026-03-07T07:09:53.157912+0000 mon.vm07 (mon.0) 1140 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:09:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:53 vm09 bash[21220]: cluster 2026-03-07T07:09:53.527704+0000 mon.vm07 (mon.0) 1141 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-07T08:09:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:53 vm09 bash[21220]: cluster 2026-03-07T07:09:53.527704+0000 mon.vm07 (mon.0) 1141 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-07T08:09:54.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:53 vm07 bash[17031]: cluster 2026-03-07T07:09:52.823665+0000 mgr.vm07.yrfcuj (mgr.14201) 840 : cluster [DBG] pgmap v496: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:54.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:53 vm07 bash[17031]: cluster 2026-03-07T07:09:52.823665+0000 mgr.vm07.yrfcuj (mgr.14201) 840 : cluster [DBG] pgmap v496: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:09:54.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:53 vm07 bash[17031]: audit 2026-03-07T07:09:52.827312+0000 mon.vm07 (mon.0) 1135 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:54.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:53 vm07 bash[17031]: audit 2026-03-07T07:09:52.827312+0000 mon.vm07 (mon.0) 1135 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:54.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:53 vm07 bash[17031]: audit 2026-03-07T07:09:52.831554+0000 mon.vm07 (mon.0) 1136 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:54.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:53 vm07 bash[17031]: audit 2026-03-07T07:09:52.831554+0000 mon.vm07 (mon.0) 1136 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:54.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:53 vm07 bash[17031]: audit 2026-03-07T07:09:53.150399+0000 mon.vm07 (mon.0) 1137 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:09:54.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:53 vm07 bash[17031]: audit 2026-03-07T07:09:53.150399+0000 mon.vm07 (mon.0) 1137 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:09:54.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:53 vm07 bash[17031]: audit 2026-03-07T07:09:53.150995+0000 mon.vm07 (mon.0) 1138 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:09:54.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:53 vm07 bash[17031]: audit 2026-03-07T07:09:53.150995+0000 mon.vm07 (mon.0) 1138 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:09:54.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:53 vm07 bash[17031]: cluster 2026-03-07T07:09:53.151923+0000 mgr.vm07.yrfcuj (mgr.14201) 841 : cluster [DBG] pgmap v497: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:09:54.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:53 vm07 bash[17031]: cluster 2026-03-07T07:09:53.151923+0000 mgr.vm07.yrfcuj (mgr.14201) 841 : cluster [DBG] pgmap v497: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:09:54.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:53 vm07 bash[17031]: audit 2026-03-07T07:09:53.156385+0000 mon.vm07 (mon.0) 1139 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:54.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:53 vm07 bash[17031]: audit 2026-03-07T07:09:53.156385+0000 mon.vm07 (mon.0) 1139 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:54.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:53 vm07 bash[17031]: audit 2026-03-07T07:09:53.157912+0000 mon.vm07 (mon.0) 1140 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:09:54.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:53 vm07 bash[17031]: audit 2026-03-07T07:09:53.157912+0000 mon.vm07 (mon.0) 1140 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:09:54.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:53 vm07 bash[17031]: cluster 2026-03-07T07:09:53.527704+0000 mon.vm07 (mon.0) 1141 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-07T08:09:54.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:53 vm07 bash[17031]: cluster 2026-03-07T07:09:53.527704+0000 mon.vm07 (mon.0) 1141 : cluster [WRN] Health check failed: 1 failed cephadm daemon(s) (CEPHADM_FAILED_DAEMON) 2026-03-07T08:09:54.910 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:09:55.072 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:09:55.072 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (6m) 5m ago 12m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:09:55.072 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 5m ago 12m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:09:55.072 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (30s) 2s ago 12m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:09:55.072 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 2s ago 12m - - 2026-03-07T08:09:55.268 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:09:55.268 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:09:55.268 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:09:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:56 vm09 bash[21220]: audit 2026-03-07T07:09:54.896440+0000 mgr.vm07.yrfcuj (mgr.14201) 842 : audit [DBG] from='client.16228 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:56 vm09 bash[21220]: audit 2026-03-07T07:09:54.896440+0000 mgr.vm07.yrfcuj (mgr.14201) 842 : audit [DBG] from='client.16228 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:56 vm09 bash[21220]: audit 2026-03-07T07:09:55.072919+0000 mgr.vm07.yrfcuj (mgr.14201) 843 : audit [DBG] from='client.16232 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:56 vm09 bash[21220]: audit 2026-03-07T07:09:55.072919+0000 mgr.vm07.yrfcuj (mgr.14201) 843 : audit [DBG] from='client.16232 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:56 vm09 bash[21220]: audit 2026-03-07T07:09:55.144785+0000 mon.vm07 (mon.0) 1142 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:56 vm09 bash[21220]: audit 2026-03-07T07:09:55.144785+0000 mon.vm07 (mon.0) 1142 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:56 vm09 bash[21220]: audit 2026-03-07T07:09:55.146061+0000 mon.vm07 (mon.0) 1143 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:09:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:56 vm09 bash[21220]: audit 2026-03-07T07:09:55.146061+0000 mon.vm07 (mon.0) 1143 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:09:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:56 vm09 bash[21220]: cluster 2026-03-07T07:09:55.153748+0000 mgr.vm07.yrfcuj (mgr.14201) 844 : cluster [DBG] pgmap v498: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:09:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:56 vm09 bash[21220]: cluster 2026-03-07T07:09:55.153748+0000 mgr.vm07.yrfcuj (mgr.14201) 844 : cluster [DBG] pgmap v498: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:09:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:56 vm09 bash[21220]: audit 2026-03-07T07:09:55.272103+0000 mon.vm07 (mon.0) 1144 : audit [DBG] from='client.? 192.168.123.107:0/1444304476' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:56 vm09 bash[21220]: audit 2026-03-07T07:09:55.272103+0000 mon.vm07 (mon.0) 1144 : audit [DBG] from='client.? 192.168.123.107:0/1444304476' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:56.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:56 vm07 bash[17031]: audit 2026-03-07T07:09:54.896440+0000 mgr.vm07.yrfcuj (mgr.14201) 842 : audit [DBG] from='client.16228 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:56.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:56 vm07 bash[17031]: audit 2026-03-07T07:09:54.896440+0000 mgr.vm07.yrfcuj (mgr.14201) 842 : audit [DBG] from='client.16228 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:56.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:56 vm07 bash[17031]: audit 2026-03-07T07:09:55.072919+0000 mgr.vm07.yrfcuj (mgr.14201) 843 : audit [DBG] from='client.16232 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:56.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:56 vm07 bash[17031]: audit 2026-03-07T07:09:55.072919+0000 mgr.vm07.yrfcuj (mgr.14201) 843 : audit [DBG] from='client.16232 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:09:56.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:56 vm07 bash[17031]: audit 2026-03-07T07:09:55.144785+0000 mon.vm07 (mon.0) 1142 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:56.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:56 vm07 bash[17031]: audit 2026-03-07T07:09:55.144785+0000 mon.vm07 (mon.0) 1142 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:09:56.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:56 vm07 bash[17031]: audit 2026-03-07T07:09:55.146061+0000 mon.vm07 (mon.0) 1143 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:09:56.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:56 vm07 bash[17031]: audit 2026-03-07T07:09:55.146061+0000 mon.vm07 (mon.0) 1143 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:09:56.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:56 vm07 bash[17031]: cluster 2026-03-07T07:09:55.153748+0000 mgr.vm07.yrfcuj (mgr.14201) 844 : cluster [DBG] pgmap v498: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:09:56.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:56 vm07 bash[17031]: cluster 2026-03-07T07:09:55.153748+0000 mgr.vm07.yrfcuj (mgr.14201) 844 : cluster [DBG] pgmap v498: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:09:56.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:56 vm07 bash[17031]: audit 2026-03-07T07:09:55.272103+0000 mon.vm07 (mon.0) 1144 : audit [DBG] from='client.? 192.168.123.107:0/1444304476' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:56.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:56 vm07 bash[17031]: audit 2026-03-07T07:09:55.272103+0000 mon.vm07 (mon.0) 1144 : audit [DBG] from='client.? 192.168.123.107:0/1444304476' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:09:58.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:58 vm09 bash[21220]: cluster 2026-03-07T07:09:57.154176+0000 mgr.vm07.yrfcuj (mgr.14201) 845 : cluster [DBG] pgmap v499: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:09:58.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:09:58 vm09 bash[21220]: cluster 2026-03-07T07:09:57.154176+0000 mgr.vm07.yrfcuj (mgr.14201) 845 : cluster [DBG] pgmap v499: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:09:58.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:58 vm07 bash[17031]: cluster 2026-03-07T07:09:57.154176+0000 mgr.vm07.yrfcuj (mgr.14201) 845 : cluster [DBG] pgmap v499: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:09:58.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:09:58 vm07 bash[17031]: cluster 2026-03-07T07:09:57.154176+0000 mgr.vm07.yrfcuj (mgr.14201) 845 : cluster [DBG] pgmap v499: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:10:00.443 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:10:00.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:00 vm07 bash[17031]: cluster 2026-03-07T07:09:59.154660+0000 mgr.vm07.yrfcuj (mgr.14201) 846 : cluster [DBG] pgmap v500: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:10:00.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:00 vm07 bash[17031]: cluster 2026-03-07T07:09:59.154660+0000 mgr.vm07.yrfcuj (mgr.14201) 846 : cluster [DBG] pgmap v500: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:10:00.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:00 vm07 bash[17031]: cluster 2026-03-07T07:10:00.000092+0000 mon.vm07 (mon.0) 1145 : cluster [WRN] overall HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:10:00.543 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:00 vm07 bash[17031]: cluster 2026-03-07T07:10:00.000092+0000 mon.vm07 (mon.0) 1145 : cluster [WRN] overall HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:10:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:00 vm09 bash[21220]: cluster 2026-03-07T07:09:59.154660+0000 mgr.vm07.yrfcuj (mgr.14201) 846 : cluster [DBG] pgmap v500: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:10:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:00 vm09 bash[21220]: cluster 2026-03-07T07:09:59.154660+0000 mgr.vm07.yrfcuj (mgr.14201) 846 : cluster [DBG] pgmap v500: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:10:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:00 vm09 bash[21220]: cluster 2026-03-07T07:10:00.000092+0000 mon.vm07 (mon.0) 1145 : cluster [WRN] overall HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:10:00.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:00 vm09 bash[21220]: cluster 2026-03-07T07:10:00.000092+0000 mon.vm07 (mon.0) 1145 : cluster [WRN] overall HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:10:00.616 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:10:00.616 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (6m) 5m ago 12m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:10:00.616 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 5m ago 12m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:10:00.616 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (36s) 7s ago 12m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:10:00.616 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 7s ago 12m - - 2026-03-07T08:10:00.807 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:10:00.807 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:10:00.807 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:10:01.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:01 vm09 bash[21220]: audit 2026-03-07T07:10:00.431130+0000 mgr.vm07.yrfcuj (mgr.14201) 847 : audit [DBG] from='client.16240 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:01.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:01 vm09 bash[21220]: audit 2026-03-07T07:10:00.431130+0000 mgr.vm07.yrfcuj (mgr.14201) 847 : audit [DBG] from='client.16240 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:01.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:01 vm09 bash[21220]: audit 2026-03-07T07:10:00.810948+0000 mon.vm07 (mon.0) 1146 : audit [DBG] from='client.? 192.168.123.107:0/20677285' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:01.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:01 vm09 bash[21220]: audit 2026-03-07T07:10:00.810948+0000 mon.vm07 (mon.0) 1146 : audit [DBG] from='client.? 192.168.123.107:0/20677285' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:01.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:01 vm07 bash[17031]: audit 2026-03-07T07:10:00.431130+0000 mgr.vm07.yrfcuj (mgr.14201) 847 : audit [DBG] from='client.16240 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:01.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:01 vm07 bash[17031]: audit 2026-03-07T07:10:00.431130+0000 mgr.vm07.yrfcuj (mgr.14201) 847 : audit [DBG] from='client.16240 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:01.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:01 vm07 bash[17031]: audit 2026-03-07T07:10:00.810948+0000 mon.vm07 (mon.0) 1146 : audit [DBG] from='client.? 192.168.123.107:0/20677285' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:01.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:01 vm07 bash[17031]: audit 2026-03-07T07:10:00.810948+0000 mon.vm07 (mon.0) 1146 : audit [DBG] from='client.? 192.168.123.107:0/20677285' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:02.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:02 vm09 bash[21220]: audit 2026-03-07T07:10:00.617564+0000 mgr.vm07.yrfcuj (mgr.14201) 848 : audit [DBG] from='client.16244 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:02.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:02 vm09 bash[21220]: audit 2026-03-07T07:10:00.617564+0000 mgr.vm07.yrfcuj (mgr.14201) 848 : audit [DBG] from='client.16244 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:02.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:02 vm09 bash[21220]: cluster 2026-03-07T07:10:01.155068+0000 mgr.vm07.yrfcuj (mgr.14201) 849 : cluster [DBG] pgmap v501: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:10:02.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:02 vm09 bash[21220]: cluster 2026-03-07T07:10:01.155068+0000 mgr.vm07.yrfcuj (mgr.14201) 849 : cluster [DBG] pgmap v501: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:10:02.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:02 vm07 bash[17031]: audit 2026-03-07T07:10:00.617564+0000 mgr.vm07.yrfcuj (mgr.14201) 848 : audit [DBG] from='client.16244 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:02.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:02 vm07 bash[17031]: audit 2026-03-07T07:10:00.617564+0000 mgr.vm07.yrfcuj (mgr.14201) 848 : audit [DBG] from='client.16244 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:02.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:02 vm07 bash[17031]: cluster 2026-03-07T07:10:01.155068+0000 mgr.vm07.yrfcuj (mgr.14201) 849 : cluster [DBG] pgmap v501: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:10:02.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:02 vm07 bash[17031]: cluster 2026-03-07T07:10:01.155068+0000 mgr.vm07.yrfcuj (mgr.14201) 849 : cluster [DBG] pgmap v501: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 99 B/s rd, 198 B/s wr, 0 op/s 2026-03-07T08:10:04.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:04 vm09 bash[21220]: cluster 2026-03-07T07:10:03.155621+0000 mgr.vm07.yrfcuj (mgr.14201) 850 : cluster [DBG] pgmap v502: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 198 B/s rd, 396 B/s wr, 0 op/s 2026-03-07T08:10:04.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:04 vm09 bash[21220]: cluster 2026-03-07T07:10:03.155621+0000 mgr.vm07.yrfcuj (mgr.14201) 850 : cluster [DBG] pgmap v502: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 198 B/s rd, 396 B/s wr, 0 op/s 2026-03-07T08:10:04.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:04 vm07 bash[17031]: cluster 2026-03-07T07:10:03.155621+0000 mgr.vm07.yrfcuj (mgr.14201) 850 : cluster [DBG] pgmap v502: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 198 B/s rd, 396 B/s wr, 0 op/s 2026-03-07T08:10:04.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:04 vm07 bash[17031]: cluster 2026-03-07T07:10:03.155621+0000 mgr.vm07.yrfcuj (mgr.14201) 850 : cluster [DBG] pgmap v502: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 198 B/s rd, 396 B/s wr, 0 op/s 2026-03-07T08:10:05.989 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:10:06.139 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:10:06.139 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (6m) 5m ago 12m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:10:06.139 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (5m) 5m ago 12m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:10:06.139 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (41s) 13s ago 12m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:10:06.139 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 13s ago 12m - - 2026-03-07T08:10:06.327 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:10:06.327 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:10:06.327 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:10:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:06 vm09 bash[21220]: cluster 2026-03-07T07:10:05.156034+0000 mgr.vm07.yrfcuj (mgr.14201) 851 : cluster [DBG] pgmap v503: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:06.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:06 vm09 bash[21220]: cluster 2026-03-07T07:10:05.156034+0000 mgr.vm07.yrfcuj (mgr.14201) 851 : cluster [DBG] pgmap v503: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:06.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:06 vm07 bash[17031]: cluster 2026-03-07T07:10:05.156034+0000 mgr.vm07.yrfcuj (mgr.14201) 851 : cluster [DBG] pgmap v503: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:06.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:06 vm07 bash[17031]: cluster 2026-03-07T07:10:05.156034+0000 mgr.vm07.yrfcuj (mgr.14201) 851 : cluster [DBG] pgmap v503: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:07 vm09 bash[21220]: audit 2026-03-07T07:10:05.976656+0000 mgr.vm07.yrfcuj (mgr.14201) 852 : audit [DBG] from='client.16252 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:07 vm09 bash[21220]: audit 2026-03-07T07:10:05.976656+0000 mgr.vm07.yrfcuj (mgr.14201) 852 : audit [DBG] from='client.16252 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:07 vm09 bash[21220]: audit 2026-03-07T07:10:06.140627+0000 mgr.vm07.yrfcuj (mgr.14201) 853 : audit [DBG] from='client.16256 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:07 vm09 bash[21220]: audit 2026-03-07T07:10:06.140627+0000 mgr.vm07.yrfcuj (mgr.14201) 853 : audit [DBG] from='client.16256 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:07 vm09 bash[21220]: audit 2026-03-07T07:10:06.330995+0000 mon.vm07 (mon.0) 1147 : audit [DBG] from='client.? 192.168.123.107:0/1673001768' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:07.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:07 vm09 bash[21220]: audit 2026-03-07T07:10:06.330995+0000 mon.vm07 (mon.0) 1147 : audit [DBG] from='client.? 192.168.123.107:0/1673001768' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:07.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:07 vm07 bash[17031]: audit 2026-03-07T07:10:05.976656+0000 mgr.vm07.yrfcuj (mgr.14201) 852 : audit [DBG] from='client.16252 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:07.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:07 vm07 bash[17031]: audit 2026-03-07T07:10:05.976656+0000 mgr.vm07.yrfcuj (mgr.14201) 852 : audit [DBG] from='client.16252 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:07.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:07 vm07 bash[17031]: audit 2026-03-07T07:10:06.140627+0000 mgr.vm07.yrfcuj (mgr.14201) 853 : audit [DBG] from='client.16256 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:07.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:07 vm07 bash[17031]: audit 2026-03-07T07:10:06.140627+0000 mgr.vm07.yrfcuj (mgr.14201) 853 : audit [DBG] from='client.16256 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:07.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:07 vm07 bash[17031]: audit 2026-03-07T07:10:06.330995+0000 mon.vm07 (mon.0) 1147 : audit [DBG] from='client.? 192.168.123.107:0/1673001768' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:07.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:07 vm07 bash[17031]: audit 2026-03-07T07:10:06.330995+0000 mon.vm07 (mon.0) 1147 : audit [DBG] from='client.? 192.168.123.107:0/1673001768' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:08.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:08 vm09 bash[21220]: cluster 2026-03-07T07:10:07.156436+0000 mgr.vm07.yrfcuj (mgr.14201) 854 : cluster [DBG] pgmap v504: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:08.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:08 vm09 bash[21220]: cluster 2026-03-07T07:10:07.156436+0000 mgr.vm07.yrfcuj (mgr.14201) 854 : cluster [DBG] pgmap v504: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:08.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:08 vm07 bash[17031]: cluster 2026-03-07T07:10:07.156436+0000 mgr.vm07.yrfcuj (mgr.14201) 854 : cluster [DBG] pgmap v504: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:08.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:08 vm07 bash[17031]: cluster 2026-03-07T07:10:07.156436+0000 mgr.vm07.yrfcuj (mgr.14201) 854 : cluster [DBG] pgmap v504: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:10 vm09 bash[21220]: cluster 2026-03-07T07:10:09.156932+0000 mgr.vm07.yrfcuj (mgr.14201) 855 : cluster [DBG] pgmap v505: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:10 vm09 bash[21220]: cluster 2026-03-07T07:10:09.156932+0000 mgr.vm07.yrfcuj (mgr.14201) 855 : cluster [DBG] pgmap v505: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:10 vm09 bash[21220]: audit 2026-03-07T07:10:10.141313+0000 mon.vm07 (mon.0) 1148 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:10:10.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:10 vm09 bash[21220]: audit 2026-03-07T07:10:10.141313+0000 mon.vm07 (mon.0) 1148 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:10:10.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:10 vm07 bash[17031]: cluster 2026-03-07T07:10:09.156932+0000 mgr.vm07.yrfcuj (mgr.14201) 855 : cluster [DBG] pgmap v505: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:10.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:10 vm07 bash[17031]: cluster 2026-03-07T07:10:09.156932+0000 mgr.vm07.yrfcuj (mgr.14201) 855 : cluster [DBG] pgmap v505: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:10.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:10 vm07 bash[17031]: audit 2026-03-07T07:10:10.141313+0000 mon.vm07 (mon.0) 1148 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:10:10.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:10 vm07 bash[17031]: audit 2026-03-07T07:10:10.141313+0000 mon.vm07 (mon.0) 1148 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:10:11.504 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:10:11.660 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:10:11.660 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (6m) 5m ago 12m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:10:11.660 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (6m) 5m ago 12m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:10:11.660 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (47s) 18s ago 12m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:10:11.660 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 18s ago 12m - - 2026-03-07T08:10:11.855 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:10:11.855 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:10:11.855 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:10:12.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:12 vm09 bash[21220]: cluster 2026-03-07T07:10:11.157386+0000 mgr.vm07.yrfcuj (mgr.14201) 856 : cluster [DBG] pgmap v506: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:12.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:12 vm09 bash[21220]: cluster 2026-03-07T07:10:11.157386+0000 mgr.vm07.yrfcuj (mgr.14201) 856 : cluster [DBG] pgmap v506: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:12.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:12 vm09 bash[21220]: audit 2026-03-07T07:10:11.491715+0000 mgr.vm07.yrfcuj (mgr.14201) 857 : audit [DBG] from='client.16264 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:12.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:12 vm09 bash[21220]: audit 2026-03-07T07:10:11.491715+0000 mgr.vm07.yrfcuj (mgr.14201) 857 : audit [DBG] from='client.16264 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:12.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:12 vm09 bash[21220]: audit 2026-03-07T07:10:11.858807+0000 mon.vm07 (mon.0) 1149 : audit [DBG] from='client.? 192.168.123.107:0/584816842' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:12.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:12 vm09 bash[21220]: audit 2026-03-07T07:10:11.858807+0000 mon.vm07 (mon.0) 1149 : audit [DBG] from='client.? 192.168.123.107:0/584816842' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:12.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:12 vm07 bash[17031]: cluster 2026-03-07T07:10:11.157386+0000 mgr.vm07.yrfcuj (mgr.14201) 856 : cluster [DBG] pgmap v506: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:12.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:12 vm07 bash[17031]: cluster 2026-03-07T07:10:11.157386+0000 mgr.vm07.yrfcuj (mgr.14201) 856 : cluster [DBG] pgmap v506: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:12.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:12 vm07 bash[17031]: audit 2026-03-07T07:10:11.491715+0000 mgr.vm07.yrfcuj (mgr.14201) 857 : audit [DBG] from='client.16264 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:12.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:12 vm07 bash[17031]: audit 2026-03-07T07:10:11.491715+0000 mgr.vm07.yrfcuj (mgr.14201) 857 : audit [DBG] from='client.16264 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:12.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:12 vm07 bash[17031]: audit 2026-03-07T07:10:11.858807+0000 mon.vm07 (mon.0) 1149 : audit [DBG] from='client.? 192.168.123.107:0/584816842' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:12.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:12 vm07 bash[17031]: audit 2026-03-07T07:10:11.858807+0000 mon.vm07 (mon.0) 1149 : audit [DBG] from='client.? 192.168.123.107:0/584816842' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:13 vm09 bash[21220]: audit 2026-03-07T07:10:11.660232+0000 mgr.vm07.yrfcuj (mgr.14201) 858 : audit [DBG] from='client.16268 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:13 vm09 bash[21220]: audit 2026-03-07T07:10:11.660232+0000 mgr.vm07.yrfcuj (mgr.14201) 858 : audit [DBG] from='client.16268 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:13.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:13 vm07 bash[17031]: audit 2026-03-07T07:10:11.660232+0000 mgr.vm07.yrfcuj (mgr.14201) 858 : audit [DBG] from='client.16268 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:13.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:13 vm07 bash[17031]: audit 2026-03-07T07:10:11.660232+0000 mgr.vm07.yrfcuj (mgr.14201) 858 : audit [DBG] from='client.16268 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:14 vm09 bash[21220]: cluster 2026-03-07T07:10:13.158241+0000 mgr.vm07.yrfcuj (mgr.14201) 859 : cluster [DBG] pgmap v507: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:14 vm09 bash[21220]: cluster 2026-03-07T07:10:13.158241+0000 mgr.vm07.yrfcuj (mgr.14201) 859 : cluster [DBG] pgmap v507: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:14.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:14 vm07 bash[17031]: cluster 2026-03-07T07:10:13.158241+0000 mgr.vm07.yrfcuj (mgr.14201) 859 : cluster [DBG] pgmap v507: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:14.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:14 vm07 bash[17031]: cluster 2026-03-07T07:10:13.158241+0000 mgr.vm07.yrfcuj (mgr.14201) 859 : cluster [DBG] pgmap v507: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:16 vm09 bash[21220]: cluster 2026-03-07T07:10:15.158797+0000 mgr.vm07.yrfcuj (mgr.14201) 860 : cluster [DBG] pgmap v508: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:16.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:16 vm09 bash[21220]: cluster 2026-03-07T07:10:15.158797+0000 mgr.vm07.yrfcuj (mgr.14201) 860 : cluster [DBG] pgmap v508: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:16.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:16 vm07 bash[17031]: cluster 2026-03-07T07:10:15.158797+0000 mgr.vm07.yrfcuj (mgr.14201) 860 : cluster [DBG] pgmap v508: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:16.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:16 vm07 bash[17031]: cluster 2026-03-07T07:10:15.158797+0000 mgr.vm07.yrfcuj (mgr.14201) 860 : cluster [DBG] pgmap v508: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:17.035 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:10:17.202 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:10:17.202 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (6m) 6m ago 12m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:10:17.202 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (6m) 6m ago 12m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:10:17.202 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (52s) 24s ago 12m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:10:17.202 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 24s ago 12m - - 2026-03-07T08:10:17.398 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:10:17.398 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:10:17.398 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:10:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:18 vm09 bash[21220]: audit 2026-03-07T07:10:17.022954+0000 mgr.vm07.yrfcuj (mgr.14201) 861 : audit [DBG] from='client.16276 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:18 vm09 bash[21220]: audit 2026-03-07T07:10:17.022954+0000 mgr.vm07.yrfcuj (mgr.14201) 861 : audit [DBG] from='client.16276 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:18 vm09 bash[21220]: cluster 2026-03-07T07:10:17.159210+0000 mgr.vm07.yrfcuj (mgr.14201) 862 : cluster [DBG] pgmap v509: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:18 vm09 bash[21220]: cluster 2026-03-07T07:10:17.159210+0000 mgr.vm07.yrfcuj (mgr.14201) 862 : cluster [DBG] pgmap v509: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:18 vm09 bash[21220]: audit 2026-03-07T07:10:17.203568+0000 mgr.vm07.yrfcuj (mgr.14201) 863 : audit [DBG] from='client.16280 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:18 vm09 bash[21220]: audit 2026-03-07T07:10:17.203568+0000 mgr.vm07.yrfcuj (mgr.14201) 863 : audit [DBG] from='client.16280 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:18 vm09 bash[21220]: audit 2026-03-07T07:10:17.401986+0000 mon.vm07 (mon.0) 1150 : audit [DBG] from='client.? 192.168.123.107:0/2868121628' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:18 vm09 bash[21220]: audit 2026-03-07T07:10:17.401986+0000 mon.vm07 (mon.0) 1150 : audit [DBG] from='client.? 192.168.123.107:0/2868121628' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:18.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:18 vm07 bash[17031]: audit 2026-03-07T07:10:17.022954+0000 mgr.vm07.yrfcuj (mgr.14201) 861 : audit [DBG] from='client.16276 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:18.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:18 vm07 bash[17031]: audit 2026-03-07T07:10:17.022954+0000 mgr.vm07.yrfcuj (mgr.14201) 861 : audit [DBG] from='client.16276 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:18.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:18 vm07 bash[17031]: cluster 2026-03-07T07:10:17.159210+0000 mgr.vm07.yrfcuj (mgr.14201) 862 : cluster [DBG] pgmap v509: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:18.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:18 vm07 bash[17031]: cluster 2026-03-07T07:10:17.159210+0000 mgr.vm07.yrfcuj (mgr.14201) 862 : cluster [DBG] pgmap v509: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:18.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:18 vm07 bash[17031]: audit 2026-03-07T07:10:17.203568+0000 mgr.vm07.yrfcuj (mgr.14201) 863 : audit [DBG] from='client.16280 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:18.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:18 vm07 bash[17031]: audit 2026-03-07T07:10:17.203568+0000 mgr.vm07.yrfcuj (mgr.14201) 863 : audit [DBG] from='client.16280 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:18.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:18 vm07 bash[17031]: audit 2026-03-07T07:10:17.401986+0000 mon.vm07 (mon.0) 1150 : audit [DBG] from='client.? 192.168.123.107:0/2868121628' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:18.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:18 vm07 bash[17031]: audit 2026-03-07T07:10:17.401986+0000 mon.vm07 (mon.0) 1150 : audit [DBG] from='client.? 192.168.123.107:0/2868121628' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:20.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:20 vm09 bash[21220]: cluster 2026-03-07T07:10:19.159596+0000 mgr.vm07.yrfcuj (mgr.14201) 864 : cluster [DBG] pgmap v510: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:20.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:20 vm09 bash[21220]: cluster 2026-03-07T07:10:19.159596+0000 mgr.vm07.yrfcuj (mgr.14201) 864 : cluster [DBG] pgmap v510: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:20.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:20 vm07 bash[17031]: cluster 2026-03-07T07:10:19.159596+0000 mgr.vm07.yrfcuj (mgr.14201) 864 : cluster [DBG] pgmap v510: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:20.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:20 vm07 bash[17031]: cluster 2026-03-07T07:10:19.159596+0000 mgr.vm07.yrfcuj (mgr.14201) 864 : cluster [DBG] pgmap v510: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:22.580 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:10:22.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:22 vm09 bash[21220]: cluster 2026-03-07T07:10:21.159958+0000 mgr.vm07.yrfcuj (mgr.14201) 865 : cluster [DBG] pgmap v511: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:22.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:22 vm09 bash[21220]: cluster 2026-03-07T07:10:21.159958+0000 mgr.vm07.yrfcuj (mgr.14201) 865 : cluster [DBG] pgmap v511: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:22.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:22 vm07 bash[17031]: cluster 2026-03-07T07:10:21.159958+0000 mgr.vm07.yrfcuj (mgr.14201) 865 : cluster [DBG] pgmap v511: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:22.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:22 vm07 bash[17031]: cluster 2026-03-07T07:10:21.159958+0000 mgr.vm07.yrfcuj (mgr.14201) 865 : cluster [DBG] pgmap v511: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:22.738 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:10:22.738 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (6m) 6m ago 12m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:10:22.738 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (6m) 6m ago 12m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:10:22.738 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (58s) 29s ago 12m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:10:22.738 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 29s ago 12m - - 2026-03-07T08:10:22.931 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:10:22.931 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:10:22.931 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:10:23.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:23 vm09 bash[21220]: audit 2026-03-07T07:10:22.568592+0000 mgr.vm07.yrfcuj (mgr.14201) 866 : audit [DBG] from='client.16288 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:23.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:23 vm09 bash[21220]: audit 2026-03-07T07:10:22.568592+0000 mgr.vm07.yrfcuj (mgr.14201) 866 : audit [DBG] from='client.16288 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:23.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:23 vm09 bash[21220]: audit 2026-03-07T07:10:22.930457+0000 mon.vm09 (mon.1) 49 : audit [DBG] from='client.? 192.168.123.107:0/172226578' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:23.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:23 vm09 bash[21220]: audit 2026-03-07T07:10:22.930457+0000 mon.vm09 (mon.1) 49 : audit [DBG] from='client.? 192.168.123.107:0/172226578' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:23.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:23 vm07 bash[17031]: audit 2026-03-07T07:10:22.568592+0000 mgr.vm07.yrfcuj (mgr.14201) 866 : audit [DBG] from='client.16288 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:23.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:23 vm07 bash[17031]: audit 2026-03-07T07:10:22.568592+0000 mgr.vm07.yrfcuj (mgr.14201) 866 : audit [DBG] from='client.16288 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:23.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:23 vm07 bash[17031]: audit 2026-03-07T07:10:22.930457+0000 mon.vm09 (mon.1) 49 : audit [DBG] from='client.? 192.168.123.107:0/172226578' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:23.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:23 vm07 bash[17031]: audit 2026-03-07T07:10:22.930457+0000 mon.vm09 (mon.1) 49 : audit [DBG] from='client.? 192.168.123.107:0/172226578' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:24 vm09 bash[21220]: audit 2026-03-07T07:10:22.740313+0000 mgr.vm07.yrfcuj (mgr.14201) 867 : audit [DBG] from='client.25507 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:24 vm09 bash[21220]: audit 2026-03-07T07:10:22.740313+0000 mgr.vm07.yrfcuj (mgr.14201) 867 : audit [DBG] from='client.25507 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:24 vm09 bash[21220]: cluster 2026-03-07T07:10:23.160487+0000 mgr.vm07.yrfcuj (mgr.14201) 868 : cluster [DBG] pgmap v512: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:24 vm09 bash[21220]: cluster 2026-03-07T07:10:23.160487+0000 mgr.vm07.yrfcuj (mgr.14201) 868 : cluster [DBG] pgmap v512: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:24.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:24 vm07 bash[17031]: audit 2026-03-07T07:10:22.740313+0000 mgr.vm07.yrfcuj (mgr.14201) 867 : audit [DBG] from='client.25507 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:24.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:24 vm07 bash[17031]: audit 2026-03-07T07:10:22.740313+0000 mgr.vm07.yrfcuj (mgr.14201) 867 : audit [DBG] from='client.25507 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:24.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:24 vm07 bash[17031]: cluster 2026-03-07T07:10:23.160487+0000 mgr.vm07.yrfcuj (mgr.14201) 868 : cluster [DBG] pgmap v512: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:24.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:24 vm07 bash[17031]: cluster 2026-03-07T07:10:23.160487+0000 mgr.vm07.yrfcuj (mgr.14201) 868 : cluster [DBG] pgmap v512: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:25 vm09 bash[21220]: audit 2026-03-07T07:10:25.141412+0000 mon.vm07 (mon.0) 1151 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:10:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:25 vm09 bash[21220]: audit 2026-03-07T07:10:25.141412+0000 mon.vm07 (mon.0) 1151 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:10:25.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:25 vm07 bash[17031]: audit 2026-03-07T07:10:25.141412+0000 mon.vm07 (mon.0) 1151 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:10:25.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:25 vm07 bash[17031]: audit 2026-03-07T07:10:25.141412+0000 mon.vm07 (mon.0) 1151 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:10:26.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:26 vm09 bash[21220]: cluster 2026-03-07T07:10:25.160925+0000 mgr.vm07.yrfcuj (mgr.14201) 869 : cluster [DBG] pgmap v513: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:10:26.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:26 vm09 bash[21220]: cluster 2026-03-07T07:10:25.160925+0000 mgr.vm07.yrfcuj (mgr.14201) 869 : cluster [DBG] pgmap v513: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:10:26.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:26 vm07 bash[17031]: cluster 2026-03-07T07:10:25.160925+0000 mgr.vm07.yrfcuj (mgr.14201) 869 : cluster [DBG] pgmap v513: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:10:26.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:26 vm07 bash[17031]: cluster 2026-03-07T07:10:25.160925+0000 mgr.vm07.yrfcuj (mgr.14201) 869 : cluster [DBG] pgmap v513: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:10:28.112 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:10:28.267 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:10:28.267 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (7m) 6m ago 12m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:10:28.267 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (6m) 6m ago 12m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:10:28.268 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (63s) 35s ago 12m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:10:28.268 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 35s ago 12m - - 2026-03-07T08:10:28.458 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:10:28.458 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:10:28.458 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:10:28.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:28 vm07 bash[17031]: cluster 2026-03-07T07:10:27.161505+0000 mgr.vm07.yrfcuj (mgr.14201) 870 : cluster [DBG] pgmap v514: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:28.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:28 vm07 bash[17031]: cluster 2026-03-07T07:10:27.161505+0000 mgr.vm07.yrfcuj (mgr.14201) 870 : cluster [DBG] pgmap v514: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:28 vm09 bash[21220]: cluster 2026-03-07T07:10:27.161505+0000 mgr.vm07.yrfcuj (mgr.14201) 870 : cluster [DBG] pgmap v514: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:28 vm09 bash[21220]: cluster 2026-03-07T07:10:27.161505+0000 mgr.vm07.yrfcuj (mgr.14201) 870 : cluster [DBG] pgmap v514: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:29.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:29 vm07 bash[17031]: audit 2026-03-07T07:10:28.099600+0000 mgr.vm07.yrfcuj (mgr.14201) 871 : audit [DBG] from='client.16300 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:29.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:29 vm07 bash[17031]: audit 2026-03-07T07:10:28.099600+0000 mgr.vm07.yrfcuj (mgr.14201) 871 : audit [DBG] from='client.16300 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:29.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:29 vm07 bash[17031]: audit 2026-03-07T07:10:28.269026+0000 mgr.vm07.yrfcuj (mgr.14201) 872 : audit [DBG] from='client.16304 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:29.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:29 vm07 bash[17031]: audit 2026-03-07T07:10:28.269026+0000 mgr.vm07.yrfcuj (mgr.14201) 872 : audit [DBG] from='client.16304 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:29.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:29 vm07 bash[17031]: audit 2026-03-07T07:10:28.462194+0000 mon.vm07 (mon.0) 1152 : audit [DBG] from='client.? 192.168.123.107:0/697282875' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:29.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:29 vm07 bash[17031]: audit 2026-03-07T07:10:28.462194+0000 mon.vm07 (mon.0) 1152 : audit [DBG] from='client.? 192.168.123.107:0/697282875' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:29.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:29 vm09 bash[21220]: audit 2026-03-07T07:10:28.099600+0000 mgr.vm07.yrfcuj (mgr.14201) 871 : audit [DBG] from='client.16300 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:29.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:29 vm09 bash[21220]: audit 2026-03-07T07:10:28.099600+0000 mgr.vm07.yrfcuj (mgr.14201) 871 : audit [DBG] from='client.16300 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:29.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:29 vm09 bash[21220]: audit 2026-03-07T07:10:28.269026+0000 mgr.vm07.yrfcuj (mgr.14201) 872 : audit [DBG] from='client.16304 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:29.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:29 vm09 bash[21220]: audit 2026-03-07T07:10:28.269026+0000 mgr.vm07.yrfcuj (mgr.14201) 872 : audit [DBG] from='client.16304 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:29.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:29 vm09 bash[21220]: audit 2026-03-07T07:10:28.462194+0000 mon.vm07 (mon.0) 1152 : audit [DBG] from='client.? 192.168.123.107:0/697282875' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:29.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:29 vm09 bash[21220]: audit 2026-03-07T07:10:28.462194+0000 mon.vm07 (mon.0) 1152 : audit [DBG] from='client.? 192.168.123.107:0/697282875' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:30.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:30 vm07 bash[17031]: cluster 2026-03-07T07:10:29.162142+0000 mgr.vm07.yrfcuj (mgr.14201) 873 : cluster [DBG] pgmap v515: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:30.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:30 vm07 bash[17031]: cluster 2026-03-07T07:10:29.162142+0000 mgr.vm07.yrfcuj (mgr.14201) 873 : cluster [DBG] pgmap v515: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:30.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:30 vm09 bash[21220]: cluster 2026-03-07T07:10:29.162142+0000 mgr.vm07.yrfcuj (mgr.14201) 873 : cluster [DBG] pgmap v515: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:30.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:30 vm09 bash[21220]: cluster 2026-03-07T07:10:29.162142+0000 mgr.vm07.yrfcuj (mgr.14201) 873 : cluster [DBG] pgmap v515: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:32.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:32 vm07 bash[17031]: cluster 2026-03-07T07:10:31.162489+0000 mgr.vm07.yrfcuj (mgr.14201) 874 : cluster [DBG] pgmap v516: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:32.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:32 vm07 bash[17031]: cluster 2026-03-07T07:10:31.162489+0000 mgr.vm07.yrfcuj (mgr.14201) 874 : cluster [DBG] pgmap v516: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:32.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:32 vm09 bash[21220]: cluster 2026-03-07T07:10:31.162489+0000 mgr.vm07.yrfcuj (mgr.14201) 874 : cluster [DBG] pgmap v516: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:32.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:32 vm09 bash[21220]: cluster 2026-03-07T07:10:31.162489+0000 mgr.vm07.yrfcuj (mgr.14201) 874 : cluster [DBG] pgmap v516: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:33.635 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:10:33.796 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:10:33.796 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (7m) 6m ago 12m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:10:33.796 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (6m) 6m ago 12m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:10:33.796 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (69s) 40s ago 12m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:10:33.796 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 40s ago 12m - - 2026-03-07T08:10:33.989 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:10:33.989 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:10:33.989 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:10:34.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:34 vm07 bash[17031]: cluster 2026-03-07T07:10:33.162946+0000 mgr.vm07.yrfcuj (mgr.14201) 875 : cluster [DBG] pgmap v517: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:34.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:34 vm07 bash[17031]: cluster 2026-03-07T07:10:33.162946+0000 mgr.vm07.yrfcuj (mgr.14201) 875 : cluster [DBG] pgmap v517: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:34.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:34 vm07 bash[17031]: audit 2026-03-07T07:10:33.993137+0000 mon.vm07 (mon.0) 1153 : audit [DBG] from='client.? 192.168.123.107:0/2313363590' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:34.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:34 vm07 bash[17031]: audit 2026-03-07T07:10:33.993137+0000 mon.vm07 (mon.0) 1153 : audit [DBG] from='client.? 192.168.123.107:0/2313363590' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:34.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:34 vm09 bash[21220]: cluster 2026-03-07T07:10:33.162946+0000 mgr.vm07.yrfcuj (mgr.14201) 875 : cluster [DBG] pgmap v517: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:34.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:34 vm09 bash[21220]: cluster 2026-03-07T07:10:33.162946+0000 mgr.vm07.yrfcuj (mgr.14201) 875 : cluster [DBG] pgmap v517: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:34.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:34 vm09 bash[21220]: audit 2026-03-07T07:10:33.993137+0000 mon.vm07 (mon.0) 1153 : audit [DBG] from='client.? 192.168.123.107:0/2313363590' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:34.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:34 vm09 bash[21220]: audit 2026-03-07T07:10:33.993137+0000 mon.vm07 (mon.0) 1153 : audit [DBG] from='client.? 192.168.123.107:0/2313363590' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:35.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:35 vm07 bash[17031]: audit 2026-03-07T07:10:33.624118+0000 mgr.vm07.yrfcuj (mgr.14201) 876 : audit [DBG] from='client.16312 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:35.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:35 vm07 bash[17031]: audit 2026-03-07T07:10:33.624118+0000 mgr.vm07.yrfcuj (mgr.14201) 876 : audit [DBG] from='client.16312 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:35.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:35 vm07 bash[17031]: audit 2026-03-07T07:10:33.797271+0000 mgr.vm07.yrfcuj (mgr.14201) 877 : audit [DBG] from='client.16316 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:35.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:35 vm07 bash[17031]: audit 2026-03-07T07:10:33.797271+0000 mgr.vm07.yrfcuj (mgr.14201) 877 : audit [DBG] from='client.16316 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:35.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:35 vm09 bash[21220]: audit 2026-03-07T07:10:33.624118+0000 mgr.vm07.yrfcuj (mgr.14201) 876 : audit [DBG] from='client.16312 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:35.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:35 vm09 bash[21220]: audit 2026-03-07T07:10:33.624118+0000 mgr.vm07.yrfcuj (mgr.14201) 876 : audit [DBG] from='client.16312 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:35.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:35 vm09 bash[21220]: audit 2026-03-07T07:10:33.797271+0000 mgr.vm07.yrfcuj (mgr.14201) 877 : audit [DBG] from='client.16316 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:35.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:35 vm09 bash[21220]: audit 2026-03-07T07:10:33.797271+0000 mgr.vm07.yrfcuj (mgr.14201) 877 : audit [DBG] from='client.16316 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:36.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:36 vm07 bash[17031]: cluster 2026-03-07T07:10:35.164762+0000 mgr.vm07.yrfcuj (mgr.14201) 878 : cluster [DBG] pgmap v518: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:36.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:36 vm07 bash[17031]: cluster 2026-03-07T07:10:35.164762+0000 mgr.vm07.yrfcuj (mgr.14201) 878 : cluster [DBG] pgmap v518: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:36 vm09 bash[21220]: cluster 2026-03-07T07:10:35.164762+0000 mgr.vm07.yrfcuj (mgr.14201) 878 : cluster [DBG] pgmap v518: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:36 vm09 bash[21220]: cluster 2026-03-07T07:10:35.164762+0000 mgr.vm07.yrfcuj (mgr.14201) 878 : cluster [DBG] pgmap v518: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:38 vm09 bash[21220]: cluster 2026-03-07T07:10:37.165217+0000 mgr.vm07.yrfcuj (mgr.14201) 879 : cluster [DBG] pgmap v519: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:38 vm09 bash[21220]: cluster 2026-03-07T07:10:37.165217+0000 mgr.vm07.yrfcuj (mgr.14201) 879 : cluster [DBG] pgmap v519: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:38.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:38 vm07 bash[17031]: cluster 2026-03-07T07:10:37.165217+0000 mgr.vm07.yrfcuj (mgr.14201) 879 : cluster [DBG] pgmap v519: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:38.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:38 vm07 bash[17031]: cluster 2026-03-07T07:10:37.165217+0000 mgr.vm07.yrfcuj (mgr.14201) 879 : cluster [DBG] pgmap v519: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:39.166 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:10:39.334 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:10:39.335 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (7m) 6m ago 13m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:10:39.335 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (6m) 6m ago 12m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:10:39.335 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (74s) 46s ago 13m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:10:39.335 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 46s ago 13m - - 2026-03-07T08:10:39.530 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:10:39.530 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:10:39.530 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:10:40.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:40 vm09 bash[21220]: audit 2026-03-07T07:10:39.153601+0000 mgr.vm07.yrfcuj (mgr.14201) 880 : audit [DBG] from='client.16324 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:40.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:40 vm09 bash[21220]: audit 2026-03-07T07:10:39.153601+0000 mgr.vm07.yrfcuj (mgr.14201) 880 : audit [DBG] from='client.16324 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:40.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:40 vm09 bash[21220]: cluster 2026-03-07T07:10:39.165614+0000 mgr.vm07.yrfcuj (mgr.14201) 881 : cluster [DBG] pgmap v520: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:40.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:40 vm09 bash[21220]: cluster 2026-03-07T07:10:39.165614+0000 mgr.vm07.yrfcuj (mgr.14201) 881 : cluster [DBG] pgmap v520: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:40.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:40 vm09 bash[21220]: audit 2026-03-07T07:10:39.336148+0000 mgr.vm07.yrfcuj (mgr.14201) 882 : audit [DBG] from='client.25533 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:40.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:40 vm09 bash[21220]: audit 2026-03-07T07:10:39.336148+0000 mgr.vm07.yrfcuj (mgr.14201) 882 : audit [DBG] from='client.25533 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:40.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:40 vm09 bash[21220]: audit 2026-03-07T07:10:39.534706+0000 mon.vm07 (mon.0) 1154 : audit [DBG] from='client.? 192.168.123.107:0/4069397338' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:40.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:40 vm09 bash[21220]: audit 2026-03-07T07:10:39.534706+0000 mon.vm07 (mon.0) 1154 : audit [DBG] from='client.? 192.168.123.107:0/4069397338' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:40.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:40 vm09 bash[21220]: audit 2026-03-07T07:10:40.141750+0000 mon.vm07 (mon.0) 1155 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:10:40.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:40 vm09 bash[21220]: audit 2026-03-07T07:10:40.141750+0000 mon.vm07 (mon.0) 1155 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:10:40.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:40 vm07 bash[17031]: audit 2026-03-07T07:10:39.153601+0000 mgr.vm07.yrfcuj (mgr.14201) 880 : audit [DBG] from='client.16324 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:40.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:40 vm07 bash[17031]: audit 2026-03-07T07:10:39.153601+0000 mgr.vm07.yrfcuj (mgr.14201) 880 : audit [DBG] from='client.16324 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:40.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:40 vm07 bash[17031]: cluster 2026-03-07T07:10:39.165614+0000 mgr.vm07.yrfcuj (mgr.14201) 881 : cluster [DBG] pgmap v520: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:40.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:40 vm07 bash[17031]: cluster 2026-03-07T07:10:39.165614+0000 mgr.vm07.yrfcuj (mgr.14201) 881 : cluster [DBG] pgmap v520: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:40.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:40 vm07 bash[17031]: audit 2026-03-07T07:10:39.336148+0000 mgr.vm07.yrfcuj (mgr.14201) 882 : audit [DBG] from='client.25533 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:40.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:40 vm07 bash[17031]: audit 2026-03-07T07:10:39.336148+0000 mgr.vm07.yrfcuj (mgr.14201) 882 : audit [DBG] from='client.25533 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:40.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:40 vm07 bash[17031]: audit 2026-03-07T07:10:39.534706+0000 mon.vm07 (mon.0) 1154 : audit [DBG] from='client.? 192.168.123.107:0/4069397338' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:40.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:40 vm07 bash[17031]: audit 2026-03-07T07:10:39.534706+0000 mon.vm07 (mon.0) 1154 : audit [DBG] from='client.? 192.168.123.107:0/4069397338' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:40.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:40 vm07 bash[17031]: audit 2026-03-07T07:10:40.141750+0000 mon.vm07 (mon.0) 1155 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:10:40.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:40 vm07 bash[17031]: audit 2026-03-07T07:10:40.141750+0000 mon.vm07 (mon.0) 1155 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:10:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:42 vm09 bash[21220]: cluster 2026-03-07T07:10:41.165998+0000 mgr.vm07.yrfcuj (mgr.14201) 883 : cluster [DBG] pgmap v521: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:42 vm09 bash[21220]: cluster 2026-03-07T07:10:41.165998+0000 mgr.vm07.yrfcuj (mgr.14201) 883 : cluster [DBG] pgmap v521: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:42.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:42 vm07 bash[17031]: cluster 2026-03-07T07:10:41.165998+0000 mgr.vm07.yrfcuj (mgr.14201) 883 : cluster [DBG] pgmap v521: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:42.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:42 vm07 bash[17031]: cluster 2026-03-07T07:10:41.165998+0000 mgr.vm07.yrfcuj (mgr.14201) 883 : cluster [DBG] pgmap v521: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:44.715 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:10:44.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:44 vm09 bash[21220]: cluster 2026-03-07T07:10:43.166558+0000 mgr.vm07.yrfcuj (mgr.14201) 884 : cluster [DBG] pgmap v522: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:44.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:44 vm09 bash[21220]: cluster 2026-03-07T07:10:43.166558+0000 mgr.vm07.yrfcuj (mgr.14201) 884 : cluster [DBG] pgmap v522: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:44.871 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:10:44.871 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (7m) 6m ago 13m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:10:44.871 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (6m) 6m ago 13m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:10:44.871 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (80s) 52s ago 13m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:10:44.871 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 52s ago 13m - - 2026-03-07T08:10:44.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:44 vm07 bash[17031]: cluster 2026-03-07T07:10:43.166558+0000 mgr.vm07.yrfcuj (mgr.14201) 884 : cluster [DBG] pgmap v522: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:44.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:44 vm07 bash[17031]: cluster 2026-03-07T07:10:43.166558+0000 mgr.vm07.yrfcuj (mgr.14201) 884 : cluster [DBG] pgmap v522: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:10:45.061 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:10:45.061 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:10:45.061 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:10:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:45 vm09 bash[21220]: audit 2026-03-07T07:10:45.064818+0000 mon.vm07 (mon.0) 1156 : audit [DBG] from='client.? 192.168.123.107:0/1726754063' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:45 vm09 bash[21220]: audit 2026-03-07T07:10:45.064818+0000 mon.vm07 (mon.0) 1156 : audit [DBG] from='client.? 192.168.123.107:0/1726754063' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:45.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:45 vm07 bash[17031]: audit 2026-03-07T07:10:45.064818+0000 mon.vm07 (mon.0) 1156 : audit [DBG] from='client.? 192.168.123.107:0/1726754063' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:45.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:45 vm07 bash[17031]: audit 2026-03-07T07:10:45.064818+0000 mon.vm07 (mon.0) 1156 : audit [DBG] from='client.? 192.168.123.107:0/1726754063' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:46.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:46 vm09 bash[21220]: audit 2026-03-07T07:10:44.701983+0000 mgr.vm07.yrfcuj (mgr.14201) 885 : audit [DBG] from='client.16336 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:46.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:46 vm09 bash[21220]: audit 2026-03-07T07:10:44.701983+0000 mgr.vm07.yrfcuj (mgr.14201) 885 : audit [DBG] from='client.16336 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:46.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:46 vm09 bash[21220]: audit 2026-03-07T07:10:44.873155+0000 mgr.vm07.yrfcuj (mgr.14201) 886 : audit [DBG] from='client.16340 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:46.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:46 vm09 bash[21220]: audit 2026-03-07T07:10:44.873155+0000 mgr.vm07.yrfcuj (mgr.14201) 886 : audit [DBG] from='client.16340 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:46.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:46 vm09 bash[21220]: cluster 2026-03-07T07:10:45.167072+0000 mgr.vm07.yrfcuj (mgr.14201) 887 : cluster [DBG] pgmap v523: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:46.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:46 vm09 bash[21220]: cluster 2026-03-07T07:10:45.167072+0000 mgr.vm07.yrfcuj (mgr.14201) 887 : cluster [DBG] pgmap v523: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:46.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:46 vm07 bash[17031]: audit 2026-03-07T07:10:44.701983+0000 mgr.vm07.yrfcuj (mgr.14201) 885 : audit [DBG] from='client.16336 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:46.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:46 vm07 bash[17031]: audit 2026-03-07T07:10:44.701983+0000 mgr.vm07.yrfcuj (mgr.14201) 885 : audit [DBG] from='client.16336 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:46.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:46 vm07 bash[17031]: audit 2026-03-07T07:10:44.873155+0000 mgr.vm07.yrfcuj (mgr.14201) 886 : audit [DBG] from='client.16340 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:46.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:46 vm07 bash[17031]: audit 2026-03-07T07:10:44.873155+0000 mgr.vm07.yrfcuj (mgr.14201) 886 : audit [DBG] from='client.16340 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:46.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:46 vm07 bash[17031]: cluster 2026-03-07T07:10:45.167072+0000 mgr.vm07.yrfcuj (mgr.14201) 887 : cluster [DBG] pgmap v523: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:46.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:46 vm07 bash[17031]: cluster 2026-03-07T07:10:45.167072+0000 mgr.vm07.yrfcuj (mgr.14201) 887 : cluster [DBG] pgmap v523: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:48 vm09 bash[21220]: cluster 2026-03-07T07:10:47.167601+0000 mgr.vm07.yrfcuj (mgr.14201) 888 : cluster [DBG] pgmap v524: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:48 vm09 bash[21220]: cluster 2026-03-07T07:10:47.167601+0000 mgr.vm07.yrfcuj (mgr.14201) 888 : cluster [DBG] pgmap v524: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:48.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:48 vm07 bash[17031]: cluster 2026-03-07T07:10:47.167601+0000 mgr.vm07.yrfcuj (mgr.14201) 888 : cluster [DBG] pgmap v524: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:48.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:48 vm07 bash[17031]: cluster 2026-03-07T07:10:47.167601+0000 mgr.vm07.yrfcuj (mgr.14201) 888 : cluster [DBG] pgmap v524: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:50.243 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:10:50.412 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:10:50.412 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (7m) 6m ago 13m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:10:50.412 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (6m) 6m ago 13m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:10:50.412 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (86s) 57s ago 13m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:10:50.412 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 57s ago 13m - - 2026-03-07T08:10:50.614 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:10:50.614 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:10:50.614 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:10:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:50 vm09 bash[21220]: cluster 2026-03-07T07:10:49.168071+0000 mgr.vm07.yrfcuj (mgr.14201) 889 : cluster [DBG] pgmap v525: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:50 vm09 bash[21220]: cluster 2026-03-07T07:10:49.168071+0000 mgr.vm07.yrfcuj (mgr.14201) 889 : cluster [DBG] pgmap v525: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:50.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:50 vm07 bash[17031]: cluster 2026-03-07T07:10:49.168071+0000 mgr.vm07.yrfcuj (mgr.14201) 889 : cluster [DBG] pgmap v525: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:50.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:50 vm07 bash[17031]: cluster 2026-03-07T07:10:49.168071+0000 mgr.vm07.yrfcuj (mgr.14201) 889 : cluster [DBG] pgmap v525: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:51 vm09 bash[21220]: audit 2026-03-07T07:10:50.231611+0000 mgr.vm07.yrfcuj (mgr.14201) 890 : audit [DBG] from='client.16348 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:51 vm09 bash[21220]: audit 2026-03-07T07:10:50.231611+0000 mgr.vm07.yrfcuj (mgr.14201) 890 : audit [DBG] from='client.16348 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:51 vm09 bash[21220]: audit 2026-03-07T07:10:50.413436+0000 mgr.vm07.yrfcuj (mgr.14201) 891 : audit [DBG] from='client.16352 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:51 vm09 bash[21220]: audit 2026-03-07T07:10:50.413436+0000 mgr.vm07.yrfcuj (mgr.14201) 891 : audit [DBG] from='client.16352 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:51 vm09 bash[21220]: audit 2026-03-07T07:10:50.618614+0000 mon.vm07 (mon.0) 1157 : audit [DBG] from='client.? 192.168.123.107:0/2398255418' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:51 vm09 bash[21220]: audit 2026-03-07T07:10:50.618614+0000 mon.vm07 (mon.0) 1157 : audit [DBG] from='client.? 192.168.123.107:0/2398255418' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:51 vm07 bash[17031]: audit 2026-03-07T07:10:50.231611+0000 mgr.vm07.yrfcuj (mgr.14201) 890 : audit [DBG] from='client.16348 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:51 vm07 bash[17031]: audit 2026-03-07T07:10:50.231611+0000 mgr.vm07.yrfcuj (mgr.14201) 890 : audit [DBG] from='client.16348 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:51 vm07 bash[17031]: audit 2026-03-07T07:10:50.413436+0000 mgr.vm07.yrfcuj (mgr.14201) 891 : audit [DBG] from='client.16352 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:51 vm07 bash[17031]: audit 2026-03-07T07:10:50.413436+0000 mgr.vm07.yrfcuj (mgr.14201) 891 : audit [DBG] from='client.16352 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:51 vm07 bash[17031]: audit 2026-03-07T07:10:50.618614+0000 mon.vm07 (mon.0) 1157 : audit [DBG] from='client.? 192.168.123.107:0/2398255418' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:51 vm07 bash[17031]: audit 2026-03-07T07:10:50.618614+0000 mon.vm07 (mon.0) 1157 : audit [DBG] from='client.? 192.168.123.107:0/2398255418' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:52 vm09 bash[21220]: cluster 2026-03-07T07:10:51.168532+0000 mgr.vm07.yrfcuj (mgr.14201) 892 : cluster [DBG] pgmap v526: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:52.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:52 vm09 bash[21220]: cluster 2026-03-07T07:10:51.168532+0000 mgr.vm07.yrfcuj (mgr.14201) 892 : cluster [DBG] pgmap v526: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:52.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:52 vm07 bash[17031]: cluster 2026-03-07T07:10:51.168532+0000 mgr.vm07.yrfcuj (mgr.14201) 892 : cluster [DBG] pgmap v526: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:52.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:52 vm07 bash[17031]: cluster 2026-03-07T07:10:51.168532+0000 mgr.vm07.yrfcuj (mgr.14201) 892 : cluster [DBG] pgmap v526: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:53.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:53 vm09 bash[21220]: audit 2026-03-07T07:10:53.202313+0000 mon.vm07 (mon.0) 1158 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:10:53.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:53 vm09 bash[21220]: audit 2026-03-07T07:10:53.202313+0000 mon.vm07 (mon.0) 1158 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:10:53.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:53 vm07 bash[17031]: audit 2026-03-07T07:10:53.202313+0000 mon.vm07 (mon.0) 1158 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:10:53.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:53 vm07 bash[17031]: audit 2026-03-07T07:10:53.202313+0000 mon.vm07 (mon.0) 1158 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:10:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:54 vm09 bash[21220]: cluster 2026-03-07T07:10:53.169127+0000 mgr.vm07.yrfcuj (mgr.14201) 893 : cluster [DBG] pgmap v527: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:54 vm09 bash[21220]: cluster 2026-03-07T07:10:53.169127+0000 mgr.vm07.yrfcuj (mgr.14201) 893 : cluster [DBG] pgmap v527: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:54 vm09 bash[21220]: audit 2026-03-07T07:10:53.542955+0000 mon.vm07 (mon.0) 1159 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:10:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:54 vm09 bash[21220]: audit 2026-03-07T07:10:53.542955+0000 mon.vm07 (mon.0) 1159 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:10:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:54 vm09 bash[21220]: audit 2026-03-07T07:10:53.543484+0000 mon.vm07 (mon.0) 1160 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:10:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:54 vm09 bash[21220]: audit 2026-03-07T07:10:53.543484+0000 mon.vm07 (mon.0) 1160 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:10:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:54 vm09 bash[21220]: cluster 2026-03-07T07:10:53.544618+0000 mgr.vm07.yrfcuj (mgr.14201) 894 : cluster [DBG] pgmap v528: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:10:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:54 vm09 bash[21220]: cluster 2026-03-07T07:10:53.544618+0000 mgr.vm07.yrfcuj (mgr.14201) 894 : cluster [DBG] pgmap v528: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:10:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:54 vm09 bash[21220]: audit 2026-03-07T07:10:53.547339+0000 mon.vm07 (mon.0) 1161 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:10:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:54 vm09 bash[21220]: audit 2026-03-07T07:10:53.547339+0000 mon.vm07 (mon.0) 1161 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:10:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:54 vm09 bash[21220]: audit 2026-03-07T07:10:53.548611+0000 mon.vm07 (mon.0) 1162 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:10:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:54 vm09 bash[21220]: audit 2026-03-07T07:10:53.548611+0000 mon.vm07 (mon.0) 1162 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:10:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:54 vm07 bash[17031]: cluster 2026-03-07T07:10:53.169127+0000 mgr.vm07.yrfcuj (mgr.14201) 893 : cluster [DBG] pgmap v527: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:54 vm07 bash[17031]: cluster 2026-03-07T07:10:53.169127+0000 mgr.vm07.yrfcuj (mgr.14201) 893 : cluster [DBG] pgmap v527: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:10:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:54 vm07 bash[17031]: audit 2026-03-07T07:10:53.542955+0000 mon.vm07 (mon.0) 1159 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:10:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:54 vm07 bash[17031]: audit 2026-03-07T07:10:53.542955+0000 mon.vm07 (mon.0) 1159 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:10:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:54 vm07 bash[17031]: audit 2026-03-07T07:10:53.543484+0000 mon.vm07 (mon.0) 1160 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:10:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:54 vm07 bash[17031]: audit 2026-03-07T07:10:53.543484+0000 mon.vm07 (mon.0) 1160 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:10:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:54 vm07 bash[17031]: cluster 2026-03-07T07:10:53.544618+0000 mgr.vm07.yrfcuj (mgr.14201) 894 : cluster [DBG] pgmap v528: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:10:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:54 vm07 bash[17031]: cluster 2026-03-07T07:10:53.544618+0000 mgr.vm07.yrfcuj (mgr.14201) 894 : cluster [DBG] pgmap v528: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:10:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:54 vm07 bash[17031]: audit 2026-03-07T07:10:53.547339+0000 mon.vm07 (mon.0) 1161 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:10:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:54 vm07 bash[17031]: audit 2026-03-07T07:10:53.547339+0000 mon.vm07 (mon.0) 1161 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:10:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:54 vm07 bash[17031]: audit 2026-03-07T07:10:53.548611+0000 mon.vm07 (mon.0) 1162 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:10:54.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:54 vm07 bash[17031]: audit 2026-03-07T07:10:53.548611+0000 mon.vm07 (mon.0) 1162 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:10:55.796 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:10:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:55 vm09 bash[21220]: audit 2026-03-07T07:10:55.142214+0000 mon.vm07 (mon.0) 1163 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:10:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:55 vm09 bash[21220]: audit 2026-03-07T07:10:55.142214+0000 mon.vm07 (mon.0) 1163 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:10:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:55 vm07 bash[17031]: audit 2026-03-07T07:10:55.142214+0000 mon.vm07 (mon.0) 1163 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:10:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:55 vm07 bash[17031]: audit 2026-03-07T07:10:55.142214+0000 mon.vm07 (mon.0) 1163 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:10:55.960 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:10:55.960 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (7m) 6m ago 13m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:10:55.960 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (6m) 6m ago 13m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:10:55.961 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (91s) 63s ago 13m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:10:55.961 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 63s ago 13m - - 2026-03-07T08:10:56.169 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:10:56.169 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:10:56.169 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:10:56.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:56 vm09 bash[21220]: cluster 2026-03-07T07:10:55.544953+0000 mgr.vm07.yrfcuj (mgr.14201) 895 : cluster [DBG] pgmap v529: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:10:56.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:56 vm09 bash[21220]: cluster 2026-03-07T07:10:55.544953+0000 mgr.vm07.yrfcuj (mgr.14201) 895 : cluster [DBG] pgmap v529: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:10:56.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:56 vm09 bash[21220]: audit 2026-03-07T07:10:56.173262+0000 mon.vm07 (mon.0) 1164 : audit [DBG] from='client.? 192.168.123.107:0/1391906065' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:56.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:56 vm09 bash[21220]: audit 2026-03-07T07:10:56.173262+0000 mon.vm07 (mon.0) 1164 : audit [DBG] from='client.? 192.168.123.107:0/1391906065' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:56.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:56 vm07 bash[17031]: cluster 2026-03-07T07:10:55.544953+0000 mgr.vm07.yrfcuj (mgr.14201) 895 : cluster [DBG] pgmap v529: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:10:56.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:56 vm07 bash[17031]: cluster 2026-03-07T07:10:55.544953+0000 mgr.vm07.yrfcuj (mgr.14201) 895 : cluster [DBG] pgmap v529: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:10:56.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:56 vm07 bash[17031]: audit 2026-03-07T07:10:56.173262+0000 mon.vm07 (mon.0) 1164 : audit [DBG] from='client.? 192.168.123.107:0/1391906065' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:56.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:56 vm07 bash[17031]: audit 2026-03-07T07:10:56.173262+0000 mon.vm07 (mon.0) 1164 : audit [DBG] from='client.? 192.168.123.107:0/1391906065' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:10:57.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:57 vm09 bash[21220]: audit 2026-03-07T07:10:55.785113+0000 mgr.vm07.yrfcuj (mgr.14201) 896 : audit [DBG] from='client.16360 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:57.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:57 vm09 bash[21220]: audit 2026-03-07T07:10:55.785113+0000 mgr.vm07.yrfcuj (mgr.14201) 896 : audit [DBG] from='client.16360 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:57.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:57 vm09 bash[21220]: audit 2026-03-07T07:10:55.962137+0000 mgr.vm07.yrfcuj (mgr.14201) 897 : audit [DBG] from='client.16364 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:57.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:57 vm09 bash[21220]: audit 2026-03-07T07:10:55.962137+0000 mgr.vm07.yrfcuj (mgr.14201) 897 : audit [DBG] from='client.16364 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:57.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:57 vm07 bash[17031]: audit 2026-03-07T07:10:55.785113+0000 mgr.vm07.yrfcuj (mgr.14201) 896 : audit [DBG] from='client.16360 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:57.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:57 vm07 bash[17031]: audit 2026-03-07T07:10:55.785113+0000 mgr.vm07.yrfcuj (mgr.14201) 896 : audit [DBG] from='client.16360 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:57.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:57 vm07 bash[17031]: audit 2026-03-07T07:10:55.962137+0000 mgr.vm07.yrfcuj (mgr.14201) 897 : audit [DBG] from='client.16364 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:57.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:57 vm07 bash[17031]: audit 2026-03-07T07:10:55.962137+0000 mgr.vm07.yrfcuj (mgr.14201) 897 : audit [DBG] from='client.16364 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:10:58.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:58 vm09 bash[21220]: cluster 2026-03-07T07:10:57.545444+0000 mgr.vm07.yrfcuj (mgr.14201) 898 : cluster [DBG] pgmap v530: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:10:58.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:10:58 vm09 bash[21220]: cluster 2026-03-07T07:10:57.545444+0000 mgr.vm07.yrfcuj (mgr.14201) 898 : cluster [DBG] pgmap v530: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:10:58.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:58 vm07 bash[17031]: cluster 2026-03-07T07:10:57.545444+0000 mgr.vm07.yrfcuj (mgr.14201) 898 : cluster [DBG] pgmap v530: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:10:58.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:10:58 vm07 bash[17031]: cluster 2026-03-07T07:10:57.545444+0000 mgr.vm07.yrfcuj (mgr.14201) 898 : cluster [DBG] pgmap v530: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:11:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:00 vm09 bash[21220]: cluster 2026-03-07T07:10:59.545929+0000 mgr.vm07.yrfcuj (mgr.14201) 899 : cluster [DBG] pgmap v531: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:11:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:00 vm09 bash[21220]: cluster 2026-03-07T07:10:59.545929+0000 mgr.vm07.yrfcuj (mgr.14201) 899 : cluster [DBG] pgmap v531: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:11:00.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:00 vm07 bash[17031]: cluster 2026-03-07T07:10:59.545929+0000 mgr.vm07.yrfcuj (mgr.14201) 899 : cluster [DBG] pgmap v531: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:11:00.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:00 vm07 bash[17031]: cluster 2026-03-07T07:10:59.545929+0000 mgr.vm07.yrfcuj (mgr.14201) 899 : cluster [DBG] pgmap v531: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:11:01.388 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:11:01.557 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:11:01.557 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (7m) 6m ago 13m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:11:01.557 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (6m) 6m ago 13m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:11:01.557 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (97s) 68s ago 13m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:11:01.557 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 68s ago 13m - - 2026-03-07T08:11:01.774 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:11:01.774 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:11:01.774 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:11:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:01 vm09 bash[21220]: audit 2026-03-07T07:11:01.364827+0000 mgr.vm07.yrfcuj (mgr.14201) 900 : audit [DBG] from='client.16372 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:01 vm09 bash[21220]: audit 2026-03-07T07:11:01.364827+0000 mgr.vm07.yrfcuj (mgr.14201) 900 : audit [DBG] from='client.16372 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:01 vm09 bash[21220]: cluster 2026-03-07T07:11:01.546344+0000 mgr.vm07.yrfcuj (mgr.14201) 901 : cluster [DBG] pgmap v532: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:11:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:01 vm09 bash[21220]: cluster 2026-03-07T07:11:01.546344+0000 mgr.vm07.yrfcuj (mgr.14201) 901 : cluster [DBG] pgmap v532: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:11:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:01 vm09 bash[21220]: audit 2026-03-07T07:11:01.558530+0000 mgr.vm07.yrfcuj (mgr.14201) 902 : audit [DBG] from='client.25559 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:01 vm09 bash[21220]: audit 2026-03-07T07:11:01.558530+0000 mgr.vm07.yrfcuj (mgr.14201) 902 : audit [DBG] from='client.25559 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:01.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:01 vm07 bash[17031]: audit 2026-03-07T07:11:01.364827+0000 mgr.vm07.yrfcuj (mgr.14201) 900 : audit [DBG] from='client.16372 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:01.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:01 vm07 bash[17031]: audit 2026-03-07T07:11:01.364827+0000 mgr.vm07.yrfcuj (mgr.14201) 900 : audit [DBG] from='client.16372 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:01.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:01 vm07 bash[17031]: cluster 2026-03-07T07:11:01.546344+0000 mgr.vm07.yrfcuj (mgr.14201) 901 : cluster [DBG] pgmap v532: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:11:01.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:01 vm07 bash[17031]: cluster 2026-03-07T07:11:01.546344+0000 mgr.vm07.yrfcuj (mgr.14201) 901 : cluster [DBG] pgmap v532: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 98 B/s rd, 197 B/s wr, 0 op/s 2026-03-07T08:11:01.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:01 vm07 bash[17031]: audit 2026-03-07T07:11:01.558530+0000 mgr.vm07.yrfcuj (mgr.14201) 902 : audit [DBG] from='client.25559 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:01.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:01 vm07 bash[17031]: audit 2026-03-07T07:11:01.558530+0000 mgr.vm07.yrfcuj (mgr.14201) 902 : audit [DBG] from='client.25559 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:02.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:02 vm07 bash[17031]: audit 2026-03-07T07:11:01.778749+0000 mon.vm07 (mon.0) 1165 : audit [DBG] from='client.? 192.168.123.107:0/3870953023' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:02.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:02 vm07 bash[17031]: audit 2026-03-07T07:11:01.778749+0000 mon.vm07 (mon.0) 1165 : audit [DBG] from='client.? 192.168.123.107:0/3870953023' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:03.099 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:02 vm09 bash[21220]: audit 2026-03-07T07:11:01.778749+0000 mon.vm07 (mon.0) 1165 : audit [DBG] from='client.? 192.168.123.107:0/3870953023' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:03.099 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:02 vm09 bash[21220]: audit 2026-03-07T07:11:01.778749+0000 mon.vm07 (mon.0) 1165 : audit [DBG] from='client.? 192.168.123.107:0/3870953023' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:04.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:04 vm09 bash[21220]: cluster 2026-03-07T07:11:03.546904+0000 mgr.vm07.yrfcuj (mgr.14201) 903 : cluster [DBG] pgmap v533: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 197 B/s rd, 394 B/s wr, 0 op/s 2026-03-07T08:11:04.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:04 vm09 bash[21220]: cluster 2026-03-07T07:11:03.546904+0000 mgr.vm07.yrfcuj (mgr.14201) 903 : cluster [DBG] pgmap v533: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 197 B/s rd, 394 B/s wr, 0 op/s 2026-03-07T08:11:04.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:04 vm07 bash[17031]: cluster 2026-03-07T07:11:03.546904+0000 mgr.vm07.yrfcuj (mgr.14201) 903 : cluster [DBG] pgmap v533: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 197 B/s rd, 394 B/s wr, 0 op/s 2026-03-07T08:11:04.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:04 vm07 bash[17031]: cluster 2026-03-07T07:11:03.546904+0000 mgr.vm07.yrfcuj (mgr.14201) 903 : cluster [DBG] pgmap v533: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 197 B/s rd, 394 B/s wr, 0 op/s 2026-03-07T08:11:06.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:06 vm07 bash[17031]: cluster 2026-03-07T07:11:05.547399+0000 mgr.vm07.yrfcuj (mgr.14201) 904 : cluster [DBG] pgmap v534: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:06.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:06 vm07 bash[17031]: cluster 2026-03-07T07:11:05.547399+0000 mgr.vm07.yrfcuj (mgr.14201) 904 : cluster [DBG] pgmap v534: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:06 vm09 bash[21220]: cluster 2026-03-07T07:11:05.547399+0000 mgr.vm07.yrfcuj (mgr.14201) 904 : cluster [DBG] pgmap v534: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:06 vm09 bash[21220]: cluster 2026-03-07T07:11:05.547399+0000 mgr.vm07.yrfcuj (mgr.14201) 904 : cluster [DBG] pgmap v534: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:06.950 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:11:07.112 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:11:07.112 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (7m) 6m ago 13m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:11:07.112 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (6m) 6m ago 13m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:11:07.112 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (102s) 74s ago 13m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:11:07.112 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 74s ago 13m - - 2026-03-07T08:11:07.321 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:11:07.321 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:11:07.321 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:11:07.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:07 vm09 bash[21220]: audit 2026-03-07T07:11:07.324742+0000 mon.vm07 (mon.0) 1166 : audit [DBG] from='client.? 192.168.123.107:0/70027649' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:07.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:07 vm09 bash[21220]: audit 2026-03-07T07:11:07.324742+0000 mon.vm07 (mon.0) 1166 : audit [DBG] from='client.? 192.168.123.107:0/70027649' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:07.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:07 vm07 bash[17031]: audit 2026-03-07T07:11:07.324742+0000 mon.vm07 (mon.0) 1166 : audit [DBG] from='client.? 192.168.123.107:0/70027649' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:07.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:07 vm07 bash[17031]: audit 2026-03-07T07:11:07.324742+0000 mon.vm07 (mon.0) 1166 : audit [DBG] from='client.? 192.168.123.107:0/70027649' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:08.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:08 vm09 bash[21220]: audit 2026-03-07T07:11:06.938173+0000 mgr.vm07.yrfcuj (mgr.14201) 905 : audit [DBG] from='client.25565 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:08.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:08 vm09 bash[21220]: audit 2026-03-07T07:11:06.938173+0000 mgr.vm07.yrfcuj (mgr.14201) 905 : audit [DBG] from='client.25565 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:08.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:08 vm09 bash[21220]: audit 2026-03-07T07:11:07.113355+0000 mgr.vm07.yrfcuj (mgr.14201) 906 : audit [DBG] from='client.16388 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:08.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:08 vm09 bash[21220]: audit 2026-03-07T07:11:07.113355+0000 mgr.vm07.yrfcuj (mgr.14201) 906 : audit [DBG] from='client.16388 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:08.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:08 vm09 bash[21220]: cluster 2026-03-07T07:11:07.547853+0000 mgr.vm07.yrfcuj (mgr.14201) 907 : cluster [DBG] pgmap v535: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:08.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:08 vm09 bash[21220]: cluster 2026-03-07T07:11:07.547853+0000 mgr.vm07.yrfcuj (mgr.14201) 907 : cluster [DBG] pgmap v535: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:08.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:08 vm07 bash[17031]: audit 2026-03-07T07:11:06.938173+0000 mgr.vm07.yrfcuj (mgr.14201) 905 : audit [DBG] from='client.25565 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:08.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:08 vm07 bash[17031]: audit 2026-03-07T07:11:06.938173+0000 mgr.vm07.yrfcuj (mgr.14201) 905 : audit [DBG] from='client.25565 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:08.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:08 vm07 bash[17031]: audit 2026-03-07T07:11:07.113355+0000 mgr.vm07.yrfcuj (mgr.14201) 906 : audit [DBG] from='client.16388 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:08.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:08 vm07 bash[17031]: audit 2026-03-07T07:11:07.113355+0000 mgr.vm07.yrfcuj (mgr.14201) 906 : audit [DBG] from='client.16388 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:08.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:08 vm07 bash[17031]: cluster 2026-03-07T07:11:07.547853+0000 mgr.vm07.yrfcuj (mgr.14201) 907 : cluster [DBG] pgmap v535: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:08.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:08 vm07 bash[17031]: cluster 2026-03-07T07:11:07.547853+0000 mgr.vm07.yrfcuj (mgr.14201) 907 : cluster [DBG] pgmap v535: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:09.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:09 vm07 bash[17031]: cluster 2026-03-07T07:11:09.548362+0000 mgr.vm07.yrfcuj (mgr.14201) 908 : cluster [DBG] pgmap v536: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:09.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:09 vm07 bash[17031]: cluster 2026-03-07T07:11:09.548362+0000 mgr.vm07.yrfcuj (mgr.14201) 908 : cluster [DBG] pgmap v536: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:10.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:09 vm09 bash[21220]: cluster 2026-03-07T07:11:09.548362+0000 mgr.vm07.yrfcuj (mgr.14201) 908 : cluster [DBG] pgmap v536: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:10.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:09 vm09 bash[21220]: cluster 2026-03-07T07:11:09.548362+0000 mgr.vm07.yrfcuj (mgr.14201) 908 : cluster [DBG] pgmap v536: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:10.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:10 vm07 bash[17031]: audit 2026-03-07T07:11:10.142352+0000 mon.vm07 (mon.0) 1167 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:11:10.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:10 vm07 bash[17031]: audit 2026-03-07T07:11:10.142352+0000 mon.vm07 (mon.0) 1167 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:11:11.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:10 vm09 bash[21220]: audit 2026-03-07T07:11:10.142352+0000 mon.vm07 (mon.0) 1167 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:11:11.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:10 vm09 bash[21220]: audit 2026-03-07T07:11:10.142352+0000 mon.vm07 (mon.0) 1167 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:11:11.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:11 vm07 bash[17031]: cluster 2026-03-07T07:11:11.548772+0000 mgr.vm07.yrfcuj (mgr.14201) 909 : cluster [DBG] pgmap v537: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:11.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:11 vm07 bash[17031]: cluster 2026-03-07T07:11:11.548772+0000 mgr.vm07.yrfcuj (mgr.14201) 909 : cluster [DBG] pgmap v537: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:12.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:11 vm09 bash[21220]: cluster 2026-03-07T07:11:11.548772+0000 mgr.vm07.yrfcuj (mgr.14201) 909 : cluster [DBG] pgmap v537: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:12.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:11 vm09 bash[21220]: cluster 2026-03-07T07:11:11.548772+0000 mgr.vm07.yrfcuj (mgr.14201) 909 : cluster [DBG] pgmap v537: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:12.512 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:11:12.678 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:11:12.678 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (7m) 7m ago 13m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:11:12.678 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (7m) 7m ago 13m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:11:12.678 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (108s) 79s ago 13m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:11:12.678 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 79s ago 13m - - 2026-03-07T08:11:12.882 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:11:12.882 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:11:12.882 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:11:12.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:12 vm07 bash[17031]: audit 2026-03-07T07:11:12.501409+0000 mgr.vm07.yrfcuj (mgr.14201) 910 : audit [DBG] from='client.16396 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:12.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:12 vm07 bash[17031]: audit 2026-03-07T07:11:12.501409+0000 mgr.vm07.yrfcuj (mgr.14201) 910 : audit [DBG] from='client.16396 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:13.099 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:12 vm09 bash[21220]: audit 2026-03-07T07:11:12.501409+0000 mgr.vm07.yrfcuj (mgr.14201) 910 : audit [DBG] from='client.16396 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:13.099 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:12 vm09 bash[21220]: audit 2026-03-07T07:11:12.501409+0000 mgr.vm07.yrfcuj (mgr.14201) 910 : audit [DBG] from='client.16396 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:13.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:13 vm07 bash[17031]: audit 2026-03-07T07:11:12.677889+0000 mgr.vm07.yrfcuj (mgr.14201) 911 : audit [DBG] from='client.16400 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:13.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:13 vm07 bash[17031]: audit 2026-03-07T07:11:12.677889+0000 mgr.vm07.yrfcuj (mgr.14201) 911 : audit [DBG] from='client.16400 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:13.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:13 vm07 bash[17031]: audit 2026-03-07T07:11:12.886389+0000 mon.vm07 (mon.0) 1168 : audit [DBG] from='client.? 192.168.123.107:0/3233903800' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:13.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:13 vm07 bash[17031]: audit 2026-03-07T07:11:12.886389+0000 mon.vm07 (mon.0) 1168 : audit [DBG] from='client.? 192.168.123.107:0/3233903800' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:13.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:13 vm07 bash[17031]: cluster 2026-03-07T07:11:13.549323+0000 mgr.vm07.yrfcuj (mgr.14201) 912 : cluster [DBG] pgmap v538: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:13.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:13 vm07 bash[17031]: cluster 2026-03-07T07:11:13.549323+0000 mgr.vm07.yrfcuj (mgr.14201) 912 : cluster [DBG] pgmap v538: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:14.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:13 vm09 bash[21220]: audit 2026-03-07T07:11:12.677889+0000 mgr.vm07.yrfcuj (mgr.14201) 911 : audit [DBG] from='client.16400 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:14.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:13 vm09 bash[21220]: audit 2026-03-07T07:11:12.677889+0000 mgr.vm07.yrfcuj (mgr.14201) 911 : audit [DBG] from='client.16400 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:14.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:13 vm09 bash[21220]: audit 2026-03-07T07:11:12.886389+0000 mon.vm07 (mon.0) 1168 : audit [DBG] from='client.? 192.168.123.107:0/3233903800' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:14.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:13 vm09 bash[21220]: audit 2026-03-07T07:11:12.886389+0000 mon.vm07 (mon.0) 1168 : audit [DBG] from='client.? 192.168.123.107:0/3233903800' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:14.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:13 vm09 bash[21220]: cluster 2026-03-07T07:11:13.549323+0000 mgr.vm07.yrfcuj (mgr.14201) 912 : cluster [DBG] pgmap v538: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:14.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:13 vm09 bash[21220]: cluster 2026-03-07T07:11:13.549323+0000 mgr.vm07.yrfcuj (mgr.14201) 912 : cluster [DBG] pgmap v538: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:16.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:16 vm09 bash[21220]: cluster 2026-03-07T07:11:15.549742+0000 mgr.vm07.yrfcuj (mgr.14201) 913 : cluster [DBG] pgmap v539: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:16.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:16 vm09 bash[21220]: cluster 2026-03-07T07:11:15.549742+0000 mgr.vm07.yrfcuj (mgr.14201) 913 : cluster [DBG] pgmap v539: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:16.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:16 vm07 bash[17031]: cluster 2026-03-07T07:11:15.549742+0000 mgr.vm07.yrfcuj (mgr.14201) 913 : cluster [DBG] pgmap v539: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:16.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:16 vm07 bash[17031]: cluster 2026-03-07T07:11:15.549742+0000 mgr.vm07.yrfcuj (mgr.14201) 913 : cluster [DBG] pgmap v539: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:17.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:17 vm07 bash[17031]: cluster 2026-03-07T07:11:17.550168+0000 mgr.vm07.yrfcuj (mgr.14201) 914 : cluster [DBG] pgmap v540: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:17.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:17 vm07 bash[17031]: cluster 2026-03-07T07:11:17.550168+0000 mgr.vm07.yrfcuj (mgr.14201) 914 : cluster [DBG] pgmap v540: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:18.069 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:11:18.099 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:17 vm09 bash[21220]: cluster 2026-03-07T07:11:17.550168+0000 mgr.vm07.yrfcuj (mgr.14201) 914 : cluster [DBG] pgmap v540: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:18.099 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:17 vm09 bash[21220]: cluster 2026-03-07T07:11:17.550168+0000 mgr.vm07.yrfcuj (mgr.14201) 914 : cluster [DBG] pgmap v540: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:18.229 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:11:18.229 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (7m) 7m ago 13m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:11:18.229 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (7m) 7m ago 13m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:11:18.229 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (113s) 85s ago 13m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:11:18.229 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 85s ago 13m - - 2026-03-07T08:11:18.436 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:11:18.436 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:11:18.436 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:11:18.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:18 vm07 bash[17031]: audit 2026-03-07T07:11:18.053030+0000 mgr.vm07.yrfcuj (mgr.14201) 915 : audit [DBG] from='client.25581 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:18.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:18 vm07 bash[17031]: audit 2026-03-07T07:11:18.053030+0000 mgr.vm07.yrfcuj (mgr.14201) 915 : audit [DBG] from='client.25581 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:18.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:18 vm07 bash[17031]: audit 2026-03-07T07:11:18.230750+0000 mgr.vm07.yrfcuj (mgr.14201) 916 : audit [DBG] from='client.16410 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:18.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:18 vm07 bash[17031]: audit 2026-03-07T07:11:18.230750+0000 mgr.vm07.yrfcuj (mgr.14201) 916 : audit [DBG] from='client.16410 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:18.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:18 vm07 bash[17031]: audit 2026-03-07T07:11:18.440392+0000 mon.vm07 (mon.0) 1169 : audit [DBG] from='client.? 192.168.123.107:0/3805272528' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:18.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:18 vm07 bash[17031]: audit 2026-03-07T07:11:18.440392+0000 mon.vm07 (mon.0) 1169 : audit [DBG] from='client.? 192.168.123.107:0/3805272528' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:19.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:18 vm09 bash[21220]: audit 2026-03-07T07:11:18.053030+0000 mgr.vm07.yrfcuj (mgr.14201) 915 : audit [DBG] from='client.25581 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:19.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:18 vm09 bash[21220]: audit 2026-03-07T07:11:18.053030+0000 mgr.vm07.yrfcuj (mgr.14201) 915 : audit [DBG] from='client.25581 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:19.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:18 vm09 bash[21220]: audit 2026-03-07T07:11:18.230750+0000 mgr.vm07.yrfcuj (mgr.14201) 916 : audit [DBG] from='client.16410 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:19.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:18 vm09 bash[21220]: audit 2026-03-07T07:11:18.230750+0000 mgr.vm07.yrfcuj (mgr.14201) 916 : audit [DBG] from='client.16410 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:19.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:18 vm09 bash[21220]: audit 2026-03-07T07:11:18.440392+0000 mon.vm07 (mon.0) 1169 : audit [DBG] from='client.? 192.168.123.107:0/3805272528' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:19.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:18 vm09 bash[21220]: audit 2026-03-07T07:11:18.440392+0000 mon.vm07 (mon.0) 1169 : audit [DBG] from='client.? 192.168.123.107:0/3805272528' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:19.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:19 vm07 bash[17031]: cluster 2026-03-07T07:11:19.550678+0000 mgr.vm07.yrfcuj (mgr.14201) 917 : cluster [DBG] pgmap v541: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:19.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:19 vm07 bash[17031]: cluster 2026-03-07T07:11:19.550678+0000 mgr.vm07.yrfcuj (mgr.14201) 917 : cluster [DBG] pgmap v541: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:20.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:19 vm09 bash[21220]: cluster 2026-03-07T07:11:19.550678+0000 mgr.vm07.yrfcuj (mgr.14201) 917 : cluster [DBG] pgmap v541: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:20.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:19 vm09 bash[21220]: cluster 2026-03-07T07:11:19.550678+0000 mgr.vm07.yrfcuj (mgr.14201) 917 : cluster [DBG] pgmap v541: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:22.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:22 vm09 bash[21220]: cluster 2026-03-07T07:11:21.551178+0000 mgr.vm07.yrfcuj (mgr.14201) 918 : cluster [DBG] pgmap v542: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:22.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:22 vm09 bash[21220]: cluster 2026-03-07T07:11:21.551178+0000 mgr.vm07.yrfcuj (mgr.14201) 918 : cluster [DBG] pgmap v542: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:22.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:22 vm07 bash[17031]: cluster 2026-03-07T07:11:21.551178+0000 mgr.vm07.yrfcuj (mgr.14201) 918 : cluster [DBG] pgmap v542: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:22.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:22 vm07 bash[17031]: cluster 2026-03-07T07:11:21.551178+0000 mgr.vm07.yrfcuj (mgr.14201) 918 : cluster [DBG] pgmap v542: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:23.625 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:11:23.804 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:11:23.804 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (7m) 7m ago 13m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:11:23.804 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (7m) 7m ago 13m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:11:23.804 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (119s) 90s ago 13m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:11:23.804 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 90s ago 13m - - 2026-03-07T08:11:24.002 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:11:24.002 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:11:24.002 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:11:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:24 vm09 bash[21220]: cluster 2026-03-07T07:11:23.551988+0000 mgr.vm07.yrfcuj (mgr.14201) 919 : cluster [DBG] pgmap v543: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:24 vm09 bash[21220]: cluster 2026-03-07T07:11:23.551988+0000 mgr.vm07.yrfcuj (mgr.14201) 919 : cluster [DBG] pgmap v543: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:24 vm09 bash[21220]: audit 2026-03-07T07:11:24.006737+0000 mon.vm07 (mon.0) 1170 : audit [DBG] from='client.? 192.168.123.107:0/1077383763' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:24 vm09 bash[21220]: audit 2026-03-07T07:11:24.006737+0000 mon.vm07 (mon.0) 1170 : audit [DBG] from='client.? 192.168.123.107:0/1077383763' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:24.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:24 vm07 bash[17031]: cluster 2026-03-07T07:11:23.551988+0000 mgr.vm07.yrfcuj (mgr.14201) 919 : cluster [DBG] pgmap v543: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:24.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:24 vm07 bash[17031]: cluster 2026-03-07T07:11:23.551988+0000 mgr.vm07.yrfcuj (mgr.14201) 919 : cluster [DBG] pgmap v543: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:24.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:24 vm07 bash[17031]: audit 2026-03-07T07:11:24.006737+0000 mon.vm07 (mon.0) 1170 : audit [DBG] from='client.? 192.168.123.107:0/1077383763' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:24.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:24 vm07 bash[17031]: audit 2026-03-07T07:11:24.006737+0000 mon.vm07 (mon.0) 1170 : audit [DBG] from='client.? 192.168.123.107:0/1077383763' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:25 vm09 bash[21220]: audit 2026-03-07T07:11:23.613535+0000 mgr.vm07.yrfcuj (mgr.14201) 920 : audit [DBG] from='client.16418 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:25 vm09 bash[21220]: audit 2026-03-07T07:11:23.613535+0000 mgr.vm07.yrfcuj (mgr.14201) 920 : audit [DBG] from='client.16418 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:25 vm09 bash[21220]: audit 2026-03-07T07:11:23.805358+0000 mgr.vm07.yrfcuj (mgr.14201) 921 : audit [DBG] from='client.16422 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:25 vm09 bash[21220]: audit 2026-03-07T07:11:23.805358+0000 mgr.vm07.yrfcuj (mgr.14201) 921 : audit [DBG] from='client.16422 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:25 vm09 bash[21220]: audit 2026-03-07T07:11:25.142817+0000 mon.vm07 (mon.0) 1171 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:11:25.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:25 vm09 bash[21220]: audit 2026-03-07T07:11:25.142817+0000 mon.vm07 (mon.0) 1171 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:11:25.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:25 vm07 bash[17031]: audit 2026-03-07T07:11:23.613535+0000 mgr.vm07.yrfcuj (mgr.14201) 920 : audit [DBG] from='client.16418 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:25.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:25 vm07 bash[17031]: audit 2026-03-07T07:11:23.613535+0000 mgr.vm07.yrfcuj (mgr.14201) 920 : audit [DBG] from='client.16418 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:25.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:25 vm07 bash[17031]: audit 2026-03-07T07:11:23.805358+0000 mgr.vm07.yrfcuj (mgr.14201) 921 : audit [DBG] from='client.16422 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:25.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:25 vm07 bash[17031]: audit 2026-03-07T07:11:23.805358+0000 mgr.vm07.yrfcuj (mgr.14201) 921 : audit [DBG] from='client.16422 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:25.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:25 vm07 bash[17031]: audit 2026-03-07T07:11:25.142817+0000 mon.vm07 (mon.0) 1171 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:11:25.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:25 vm07 bash[17031]: audit 2026-03-07T07:11:25.142817+0000 mon.vm07 (mon.0) 1171 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:11:26.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:26 vm09 bash[21220]: cluster 2026-03-07T07:11:25.552523+0000 mgr.vm07.yrfcuj (mgr.14201) 922 : cluster [DBG] pgmap v544: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:11:26.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:26 vm09 bash[21220]: cluster 2026-03-07T07:11:25.552523+0000 mgr.vm07.yrfcuj (mgr.14201) 922 : cluster [DBG] pgmap v544: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:11:26.643 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:26 vm07 bash[17031]: cluster 2026-03-07T07:11:25.552523+0000 mgr.vm07.yrfcuj (mgr.14201) 922 : cluster [DBG] pgmap v544: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:11:26.644 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:26 vm07 bash[17031]: cluster 2026-03-07T07:11:25.552523+0000 mgr.vm07.yrfcuj (mgr.14201) 922 : cluster [DBG] pgmap v544: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:11:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:28 vm09 bash[21220]: cluster 2026-03-07T07:11:27.552908+0000 mgr.vm07.yrfcuj (mgr.14201) 923 : cluster [DBG] pgmap v545: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:28 vm09 bash[21220]: cluster 2026-03-07T07:11:27.552908+0000 mgr.vm07.yrfcuj (mgr.14201) 923 : cluster [DBG] pgmap v545: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:28.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:28 vm07 bash[17031]: cluster 2026-03-07T07:11:27.552908+0000 mgr.vm07.yrfcuj (mgr.14201) 923 : cluster [DBG] pgmap v545: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:28.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:28 vm07 bash[17031]: cluster 2026-03-07T07:11:27.552908+0000 mgr.vm07.yrfcuj (mgr.14201) 923 : cluster [DBG] pgmap v545: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:29.194 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:11:29.373 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:11:29.373 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (8m) 7m ago 13m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:11:29.373 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (7m) 7m ago 13m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:11:29.373 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 96s ago 13m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:11:29.373 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 96s ago 13m - - 2026-03-07T08:11:29.572 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:11:29.572 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:11:29.572 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:11:29.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:29 vm07 bash[17031]: audit 2026-03-07T07:11:29.576163+0000 mon.vm07 (mon.0) 1172 : audit [DBG] from='client.? 192.168.123.107:0/377754096' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:29.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:29 vm07 bash[17031]: audit 2026-03-07T07:11:29.576163+0000 mon.vm07 (mon.0) 1172 : audit [DBG] from='client.? 192.168.123.107:0/377754096' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:29 vm09 bash[21220]: audit 2026-03-07T07:11:29.576163+0000 mon.vm07 (mon.0) 1172 : audit [DBG] from='client.? 192.168.123.107:0/377754096' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:30.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:29 vm09 bash[21220]: audit 2026-03-07T07:11:29.576163+0000 mon.vm07 (mon.0) 1172 : audit [DBG] from='client.? 192.168.123.107:0/377754096' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:30.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:30 vm07 bash[17031]: audit 2026-03-07T07:11:29.181380+0000 mgr.vm07.yrfcuj (mgr.14201) 924 : audit [DBG] from='client.16430 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:30.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:30 vm07 bash[17031]: audit 2026-03-07T07:11:29.181380+0000 mgr.vm07.yrfcuj (mgr.14201) 924 : audit [DBG] from='client.16430 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:30.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:30 vm07 bash[17031]: audit 2026-03-07T07:11:29.374912+0000 mgr.vm07.yrfcuj (mgr.14201) 925 : audit [DBG] from='client.16434 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:30.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:30 vm07 bash[17031]: audit 2026-03-07T07:11:29.374912+0000 mgr.vm07.yrfcuj (mgr.14201) 925 : audit [DBG] from='client.16434 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:30.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:30 vm07 bash[17031]: cluster 2026-03-07T07:11:29.553367+0000 mgr.vm07.yrfcuj (mgr.14201) 926 : cluster [DBG] pgmap v546: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:30.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:30 vm07 bash[17031]: cluster 2026-03-07T07:11:29.553367+0000 mgr.vm07.yrfcuj (mgr.14201) 926 : cluster [DBG] pgmap v546: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:31.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:30 vm09 bash[21220]: audit 2026-03-07T07:11:29.181380+0000 mgr.vm07.yrfcuj (mgr.14201) 924 : audit [DBG] from='client.16430 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:31.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:30 vm09 bash[21220]: audit 2026-03-07T07:11:29.181380+0000 mgr.vm07.yrfcuj (mgr.14201) 924 : audit [DBG] from='client.16430 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:31.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:30 vm09 bash[21220]: audit 2026-03-07T07:11:29.374912+0000 mgr.vm07.yrfcuj (mgr.14201) 925 : audit [DBG] from='client.16434 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:31.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:30 vm09 bash[21220]: audit 2026-03-07T07:11:29.374912+0000 mgr.vm07.yrfcuj (mgr.14201) 925 : audit [DBG] from='client.16434 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:31.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:30 vm09 bash[21220]: cluster 2026-03-07T07:11:29.553367+0000 mgr.vm07.yrfcuj (mgr.14201) 926 : cluster [DBG] pgmap v546: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:31.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:30 vm09 bash[21220]: cluster 2026-03-07T07:11:29.553367+0000 mgr.vm07.yrfcuj (mgr.14201) 926 : cluster [DBG] pgmap v546: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:31.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:31 vm07 bash[17031]: cluster 2026-03-07T07:11:31.553865+0000 mgr.vm07.yrfcuj (mgr.14201) 927 : cluster [DBG] pgmap v547: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:31.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:31 vm07 bash[17031]: cluster 2026-03-07T07:11:31.553865+0000 mgr.vm07.yrfcuj (mgr.14201) 927 : cluster [DBG] pgmap v547: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:32.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:31 vm09 bash[21220]: cluster 2026-03-07T07:11:31.553865+0000 mgr.vm07.yrfcuj (mgr.14201) 927 : cluster [DBG] pgmap v547: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:32.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:31 vm09 bash[21220]: cluster 2026-03-07T07:11:31.553865+0000 mgr.vm07.yrfcuj (mgr.14201) 927 : cluster [DBG] pgmap v547: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:33 vm09 bash[21220]: cluster 2026-03-07T07:11:33.554384+0000 mgr.vm07.yrfcuj (mgr.14201) 928 : cluster [DBG] pgmap v548: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:33 vm09 bash[21220]: cluster 2026-03-07T07:11:33.554384+0000 mgr.vm07.yrfcuj (mgr.14201) 928 : cluster [DBG] pgmap v548: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:34.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:33 vm07 bash[17031]: cluster 2026-03-07T07:11:33.554384+0000 mgr.vm07.yrfcuj (mgr.14201) 928 : cluster [DBG] pgmap v548: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:34.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:33 vm07 bash[17031]: cluster 2026-03-07T07:11:33.554384+0000 mgr.vm07.yrfcuj (mgr.14201) 928 : cluster [DBG] pgmap v548: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:34.769 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:11:34.928 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:11:34.928 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (8m) 7m ago 13m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:11:34.928 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (7m) 7m ago 13m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:11:34.928 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 102s ago 13m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:11:34.929 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 102s ago 13m - - 2026-03-07T08:11:35.141 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:11:35.142 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:11:35.142 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:11:35.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:35 vm09 bash[21220]: audit 2026-03-07T07:11:35.145829+0000 mon.vm07 (mon.0) 1173 : audit [DBG] from='client.? 192.168.123.107:0/1967658342' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:35.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:35 vm09 bash[21220]: audit 2026-03-07T07:11:35.145829+0000 mon.vm07 (mon.0) 1173 : audit [DBG] from='client.? 192.168.123.107:0/1967658342' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:35.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:35 vm07 bash[17031]: audit 2026-03-07T07:11:35.145829+0000 mon.vm07 (mon.0) 1173 : audit [DBG] from='client.? 192.168.123.107:0/1967658342' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:35.649 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:35 vm07 bash[17031]: audit 2026-03-07T07:11:35.145829+0000 mon.vm07 (mon.0) 1173 : audit [DBG] from='client.? 192.168.123.107:0/1967658342' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:36.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:36 vm09 bash[21220]: audit 2026-03-07T07:11:34.756168+0000 mgr.vm07.yrfcuj (mgr.14201) 929 : audit [DBG] from='client.16442 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:36.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:36 vm09 bash[21220]: audit 2026-03-07T07:11:34.756168+0000 mgr.vm07.yrfcuj (mgr.14201) 929 : audit [DBG] from='client.16442 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:36.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:36 vm09 bash[21220]: audit 2026-03-07T07:11:34.927995+0000 mgr.vm07.yrfcuj (mgr.14201) 930 : audit [DBG] from='client.16446 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:36.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:36 vm09 bash[21220]: audit 2026-03-07T07:11:34.927995+0000 mgr.vm07.yrfcuj (mgr.14201) 930 : audit [DBG] from='client.16446 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:36.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:36 vm09 bash[21220]: cluster 2026-03-07T07:11:35.554830+0000 mgr.vm07.yrfcuj (mgr.14201) 931 : cluster [DBG] pgmap v549: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:36.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:36 vm09 bash[21220]: cluster 2026-03-07T07:11:35.554830+0000 mgr.vm07.yrfcuj (mgr.14201) 931 : cluster [DBG] pgmap v549: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:36.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:36 vm07 bash[17031]: audit 2026-03-07T07:11:34.756168+0000 mgr.vm07.yrfcuj (mgr.14201) 929 : audit [DBG] from='client.16442 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:36.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:36 vm07 bash[17031]: audit 2026-03-07T07:11:34.756168+0000 mgr.vm07.yrfcuj (mgr.14201) 929 : audit [DBG] from='client.16442 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:36.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:36 vm07 bash[17031]: audit 2026-03-07T07:11:34.927995+0000 mgr.vm07.yrfcuj (mgr.14201) 930 : audit [DBG] from='client.16446 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:36.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:36 vm07 bash[17031]: audit 2026-03-07T07:11:34.927995+0000 mgr.vm07.yrfcuj (mgr.14201) 930 : audit [DBG] from='client.16446 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:36.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:36 vm07 bash[17031]: cluster 2026-03-07T07:11:35.554830+0000 mgr.vm07.yrfcuj (mgr.14201) 931 : cluster [DBG] pgmap v549: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:36.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:36 vm07 bash[17031]: cluster 2026-03-07T07:11:35.554830+0000 mgr.vm07.yrfcuj (mgr.14201) 931 : cluster [DBG] pgmap v549: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:38 vm09 bash[21220]: cluster 2026-03-07T07:11:37.555412+0000 mgr.vm07.yrfcuj (mgr.14201) 932 : cluster [DBG] pgmap v550: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:38.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:38 vm09 bash[21220]: cluster 2026-03-07T07:11:37.555412+0000 mgr.vm07.yrfcuj (mgr.14201) 932 : cluster [DBG] pgmap v550: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:38.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:38 vm07 bash[17031]: cluster 2026-03-07T07:11:37.555412+0000 mgr.vm07.yrfcuj (mgr.14201) 932 : cluster [DBG] pgmap v550: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:38.899 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:38 vm07 bash[17031]: cluster 2026-03-07T07:11:37.555412+0000 mgr.vm07.yrfcuj (mgr.14201) 932 : cluster [DBG] pgmap v550: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:39 vm09 bash[21220]: cluster 2026-03-07T07:11:39.555898+0000 mgr.vm07.yrfcuj (mgr.14201) 933 : cluster [DBG] pgmap v551: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:39 vm09 bash[21220]: cluster 2026-03-07T07:11:39.555898+0000 mgr.vm07.yrfcuj (mgr.14201) 933 : cluster [DBG] pgmap v551: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:40.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:39 vm07 bash[17031]: cluster 2026-03-07T07:11:39.555898+0000 mgr.vm07.yrfcuj (mgr.14201) 933 : cluster [DBG] pgmap v551: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:40.149 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:39 vm07 bash[17031]: cluster 2026-03-07T07:11:39.555898+0000 mgr.vm07.yrfcuj (mgr.14201) 933 : cluster [DBG] pgmap v551: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:40.336 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:11:40.499 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:11:40.503 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (8m) 7m ago 14m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:11:40.503 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (7m) 7m ago 14m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:11:40.503 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 107s ago 14m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:11:40.503 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 107s ago 14m - - 2026-03-07T08:11:40.787 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:11:40.787 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:11:40.787 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:11:41.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:41 vm07 bash[17031]: audit 2026-03-07T07:11:40.142997+0000 mon.vm07 (mon.0) 1174 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:11:41.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:41 vm07 bash[17031]: audit 2026-03-07T07:11:40.142997+0000 mon.vm07 (mon.0) 1174 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:11:41.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:41 vm07 bash[17031]: audit 2026-03-07T07:11:40.324197+0000 mgr.vm07.yrfcuj (mgr.14201) 934 : audit [DBG] from='client.16454 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:41.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:41 vm07 bash[17031]: audit 2026-03-07T07:11:40.324197+0000 mgr.vm07.yrfcuj (mgr.14201) 934 : audit [DBG] from='client.16454 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:41.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:41 vm07 bash[17031]: audit 2026-03-07T07:11:40.500379+0000 mgr.vm07.yrfcuj (mgr.14201) 935 : audit [DBG] from='client.16458 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:41.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:41 vm07 bash[17031]: audit 2026-03-07T07:11:40.500379+0000 mgr.vm07.yrfcuj (mgr.14201) 935 : audit [DBG] from='client.16458 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:41 vm09 bash[21220]: audit 2026-03-07T07:11:40.142997+0000 mon.vm07 (mon.0) 1174 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:11:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:41 vm09 bash[21220]: audit 2026-03-07T07:11:40.142997+0000 mon.vm07 (mon.0) 1174 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:11:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:41 vm09 bash[21220]: audit 2026-03-07T07:11:40.324197+0000 mgr.vm07.yrfcuj (mgr.14201) 934 : audit [DBG] from='client.16454 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:41 vm09 bash[21220]: audit 2026-03-07T07:11:40.324197+0000 mgr.vm07.yrfcuj (mgr.14201) 934 : audit [DBG] from='client.16454 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:41 vm09 bash[21220]: audit 2026-03-07T07:11:40.500379+0000 mgr.vm07.yrfcuj (mgr.14201) 935 : audit [DBG] from='client.16458 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:41.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:41 vm09 bash[21220]: audit 2026-03-07T07:11:40.500379+0000 mgr.vm07.yrfcuj (mgr.14201) 935 : audit [DBG] from='client.16458 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:42.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:42 vm07 bash[17031]: audit 2026-03-07T07:11:40.791275+0000 mon.vm07 (mon.0) 1175 : audit [DBG] from='client.? 192.168.123.107:0/3525752512' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:42.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:42 vm07 bash[17031]: audit 2026-03-07T07:11:40.791275+0000 mon.vm07 (mon.0) 1175 : audit [DBG] from='client.? 192.168.123.107:0/3525752512' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:42.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:42 vm07 bash[17031]: cluster 2026-03-07T07:11:41.556311+0000 mgr.vm07.yrfcuj (mgr.14201) 936 : cluster [DBG] pgmap v552: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:42.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:42 vm07 bash[17031]: cluster 2026-03-07T07:11:41.556311+0000 mgr.vm07.yrfcuj (mgr.14201) 936 : cluster [DBG] pgmap v552: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:42.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:42 vm09 bash[21220]: audit 2026-03-07T07:11:40.791275+0000 mon.vm07 (mon.0) 1175 : audit [DBG] from='client.? 192.168.123.107:0/3525752512' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:42.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:42 vm09 bash[21220]: audit 2026-03-07T07:11:40.791275+0000 mon.vm07 (mon.0) 1175 : audit [DBG] from='client.? 192.168.123.107:0/3525752512' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:42.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:42 vm09 bash[21220]: cluster 2026-03-07T07:11:41.556311+0000 mgr.vm07.yrfcuj (mgr.14201) 936 : cluster [DBG] pgmap v552: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:42.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:42 vm09 bash[21220]: cluster 2026-03-07T07:11:41.556311+0000 mgr.vm07.yrfcuj (mgr.14201) 936 : cluster [DBG] pgmap v552: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:44.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:44 vm09 bash[21220]: cluster 2026-03-07T07:11:43.556873+0000 mgr.vm07.yrfcuj (mgr.14201) 937 : cluster [DBG] pgmap v553: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:44.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:44 vm09 bash[21220]: cluster 2026-03-07T07:11:43.556873+0000 mgr.vm07.yrfcuj (mgr.14201) 937 : cluster [DBG] pgmap v553: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:44.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:44 vm07 bash[17031]: cluster 2026-03-07T07:11:43.556873+0000 mgr.vm07.yrfcuj (mgr.14201) 937 : cluster [DBG] pgmap v553: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:44.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:44 vm07 bash[17031]: cluster 2026-03-07T07:11:43.556873+0000 mgr.vm07.yrfcuj (mgr.14201) 937 : cluster [DBG] pgmap v553: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:11:45.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:45 vm07 bash[17031]: cluster 2026-03-07T07:11:45.557469+0000 mgr.vm07.yrfcuj (mgr.14201) 938 : cluster [DBG] pgmap v554: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:45.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:45 vm07 bash[17031]: cluster 2026-03-07T07:11:45.557469+0000 mgr.vm07.yrfcuj (mgr.14201) 938 : cluster [DBG] pgmap v554: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:45.974 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:11:46.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:45 vm09 bash[21220]: cluster 2026-03-07T07:11:45.557469+0000 mgr.vm07.yrfcuj (mgr.14201) 938 : cluster [DBG] pgmap v554: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:46.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:45 vm09 bash[21220]: cluster 2026-03-07T07:11:45.557469+0000 mgr.vm07.yrfcuj (mgr.14201) 938 : cluster [DBG] pgmap v554: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:46.130 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:11:46.131 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (8m) 7m ago 14m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:11:46.131 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (7m) 7m ago 14m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:11:46.131 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 113s ago 14m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:11:46.131 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 113s ago 14m - - 2026-03-07T08:11:46.324 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:11:46.325 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:11:46.325 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:11:46.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:46 vm07 bash[17031]: audit 2026-03-07T07:11:45.960853+0000 mgr.vm07.yrfcuj (mgr.14201) 939 : audit [DBG] from='client.25617 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:46.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:46 vm07 bash[17031]: audit 2026-03-07T07:11:45.960853+0000 mgr.vm07.yrfcuj (mgr.14201) 939 : audit [DBG] from='client.25617 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:46.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:46 vm07 bash[17031]: audit 2026-03-07T07:11:46.131998+0000 mgr.vm07.yrfcuj (mgr.14201) 940 : audit [DBG] from='client.25621 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:46.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:46 vm07 bash[17031]: audit 2026-03-07T07:11:46.131998+0000 mgr.vm07.yrfcuj (mgr.14201) 940 : audit [DBG] from='client.25621 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:46.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:46 vm07 bash[17031]: audit 2026-03-07T07:11:46.323622+0000 mon.vm09 (mon.1) 50 : audit [DBG] from='client.? 192.168.123.107:0/3179046597' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:46.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:46 vm07 bash[17031]: audit 2026-03-07T07:11:46.323622+0000 mon.vm09 (mon.1) 50 : audit [DBG] from='client.? 192.168.123.107:0/3179046597' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:47.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:46 vm09 bash[21220]: audit 2026-03-07T07:11:45.960853+0000 mgr.vm07.yrfcuj (mgr.14201) 939 : audit [DBG] from='client.25617 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:47.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:46 vm09 bash[21220]: audit 2026-03-07T07:11:45.960853+0000 mgr.vm07.yrfcuj (mgr.14201) 939 : audit [DBG] from='client.25617 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:47.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:46 vm09 bash[21220]: audit 2026-03-07T07:11:46.131998+0000 mgr.vm07.yrfcuj (mgr.14201) 940 : audit [DBG] from='client.25621 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:47.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:46 vm09 bash[21220]: audit 2026-03-07T07:11:46.131998+0000 mgr.vm07.yrfcuj (mgr.14201) 940 : audit [DBG] from='client.25621 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:47.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:46 vm09 bash[21220]: audit 2026-03-07T07:11:46.323622+0000 mon.vm09 (mon.1) 50 : audit [DBG] from='client.? 192.168.123.107:0/3179046597' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:47.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:46 vm09 bash[21220]: audit 2026-03-07T07:11:46.323622+0000 mon.vm09 (mon.1) 50 : audit [DBG] from='client.? 192.168.123.107:0/3179046597' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:47.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:47 vm07 bash[17031]: cluster 2026-03-07T07:11:47.557902+0000 mgr.vm07.yrfcuj (mgr.14201) 941 : cluster [DBG] pgmap v555: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:47.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:47 vm07 bash[17031]: cluster 2026-03-07T07:11:47.557902+0000 mgr.vm07.yrfcuj (mgr.14201) 941 : cluster [DBG] pgmap v555: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:48.100 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:47 vm09 bash[21220]: cluster 2026-03-07T07:11:47.557902+0000 mgr.vm07.yrfcuj (mgr.14201) 941 : cluster [DBG] pgmap v555: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:48.101 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:47 vm09 bash[21220]: cluster 2026-03-07T07:11:47.557902+0000 mgr.vm07.yrfcuj (mgr.14201) 941 : cluster [DBG] pgmap v555: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:50.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:50 vm07 bash[17031]: cluster 2026-03-07T07:11:49.558385+0000 mgr.vm07.yrfcuj (mgr.14201) 942 : cluster [DBG] pgmap v556: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:50.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:50 vm07 bash[17031]: cluster 2026-03-07T07:11:49.558385+0000 mgr.vm07.yrfcuj (mgr.14201) 942 : cluster [DBG] pgmap v556: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:51.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:50 vm09 bash[21220]: cluster 2026-03-07T07:11:49.558385+0000 mgr.vm07.yrfcuj (mgr.14201) 942 : cluster [DBG] pgmap v556: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:51.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:50 vm09 bash[21220]: cluster 2026-03-07T07:11:49.558385+0000 mgr.vm07.yrfcuj (mgr.14201) 942 : cluster [DBG] pgmap v556: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:51.512 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:11:51.678 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:11:51.678 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (8m) 7m ago 14m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:11:51.678 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (7m) 7m ago 14m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:11:51.678 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 118s ago 14m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:11:51.679 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 118s ago 14m - - 2026-03-07T08:11:51.873 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:11:51.873 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:11:51.873 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:11:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:51 vm07 bash[17031]: audit 2026-03-07T07:11:51.500599+0000 mgr.vm07.yrfcuj (mgr.14201) 943 : audit [DBG] from='client.16478 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:51 vm07 bash[17031]: audit 2026-03-07T07:11:51.500599+0000 mgr.vm07.yrfcuj (mgr.14201) 943 : audit [DBG] from='client.16478 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:51 vm07 bash[17031]: cluster 2026-03-07T07:11:51.558751+0000 mgr.vm07.yrfcuj (mgr.14201) 944 : cluster [DBG] pgmap v557: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:51 vm07 bash[17031]: cluster 2026-03-07T07:11:51.558751+0000 mgr.vm07.yrfcuj (mgr.14201) 944 : cluster [DBG] pgmap v557: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:52.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:51 vm09 bash[21220]: audit 2026-03-07T07:11:51.500599+0000 mgr.vm07.yrfcuj (mgr.14201) 943 : audit [DBG] from='client.16478 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:52.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:51 vm09 bash[21220]: audit 2026-03-07T07:11:51.500599+0000 mgr.vm07.yrfcuj (mgr.14201) 943 : audit [DBG] from='client.16478 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:52.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:51 vm09 bash[21220]: cluster 2026-03-07T07:11:51.558751+0000 mgr.vm07.yrfcuj (mgr.14201) 944 : cluster [DBG] pgmap v557: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:52.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:51 vm09 bash[21220]: cluster 2026-03-07T07:11:51.558751+0000 mgr.vm07.yrfcuj (mgr.14201) 944 : cluster [DBG] pgmap v557: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:52.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:52 vm07 bash[17031]: audit 2026-03-07T07:11:51.679904+0000 mgr.vm07.yrfcuj (mgr.14201) 945 : audit [DBG] from='client.16482 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:52.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:52 vm07 bash[17031]: audit 2026-03-07T07:11:51.679904+0000 mgr.vm07.yrfcuj (mgr.14201) 945 : audit [DBG] from='client.16482 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:52.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:52 vm07 bash[17031]: audit 2026-03-07T07:11:51.877780+0000 mon.vm07 (mon.0) 1176 : audit [DBG] from='client.? 192.168.123.107:0/464234447' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:52.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:52 vm07 bash[17031]: audit 2026-03-07T07:11:51.877780+0000 mon.vm07 (mon.0) 1176 : audit [DBG] from='client.? 192.168.123.107:0/464234447' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:53.101 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:52 vm09 bash[21220]: audit 2026-03-07T07:11:51.679904+0000 mgr.vm07.yrfcuj (mgr.14201) 945 : audit [DBG] from='client.16482 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:53.101 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:52 vm09 bash[21220]: audit 2026-03-07T07:11:51.679904+0000 mgr.vm07.yrfcuj (mgr.14201) 945 : audit [DBG] from='client.16482 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:53.101 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:52 vm09 bash[21220]: audit 2026-03-07T07:11:51.877780+0000 mon.vm07 (mon.0) 1176 : audit [DBG] from='client.? 192.168.123.107:0/464234447' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:53.101 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:52 vm09 bash[21220]: audit 2026-03-07T07:11:51.877780+0000 mon.vm07 (mon.0) 1176 : audit [DBG] from='client.? 192.168.123.107:0/464234447' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:53.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:53 vm07 bash[17031]: cluster 2026-03-07T07:11:53.559221+0000 mgr.vm07.yrfcuj (mgr.14201) 946 : cluster [DBG] pgmap v558: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:53.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:53 vm07 bash[17031]: cluster 2026-03-07T07:11:53.559221+0000 mgr.vm07.yrfcuj (mgr.14201) 946 : cluster [DBG] pgmap v558: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:53.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:53 vm07 bash[17031]: audit 2026-03-07T07:11:53.594799+0000 mon.vm07 (mon.0) 1177 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:11:53.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:53 vm07 bash[17031]: audit 2026-03-07T07:11:53.594799+0000 mon.vm07 (mon.0) 1177 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:11:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:53 vm09 bash[21220]: cluster 2026-03-07T07:11:53.559221+0000 mgr.vm07.yrfcuj (mgr.14201) 946 : cluster [DBG] pgmap v558: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:53 vm09 bash[21220]: cluster 2026-03-07T07:11:53.559221+0000 mgr.vm07.yrfcuj (mgr.14201) 946 : cluster [DBG] pgmap v558: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:11:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:53 vm09 bash[21220]: audit 2026-03-07T07:11:53.594799+0000 mon.vm07 (mon.0) 1177 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:11:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:53 vm09 bash[21220]: audit 2026-03-07T07:11:53.594799+0000 mon.vm07 (mon.0) 1177 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:11:55.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:54 vm09 bash[21220]: audit 2026-03-07T07:11:53.936087+0000 mon.vm07 (mon.0) 1178 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:11:55.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:54 vm09 bash[21220]: audit 2026-03-07T07:11:53.936087+0000 mon.vm07 (mon.0) 1178 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:11:55.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:54 vm09 bash[21220]: audit 2026-03-07T07:11:53.937152+0000 mon.vm07 (mon.0) 1179 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:11:55.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:54 vm09 bash[21220]: audit 2026-03-07T07:11:53.937152+0000 mon.vm07 (mon.0) 1179 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:11:55.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:54 vm09 bash[21220]: cluster 2026-03-07T07:11:53.938834+0000 mgr.vm07.yrfcuj (mgr.14201) 947 : cluster [DBG] pgmap v559: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:11:55.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:54 vm09 bash[21220]: cluster 2026-03-07T07:11:53.938834+0000 mgr.vm07.yrfcuj (mgr.14201) 947 : cluster [DBG] pgmap v559: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:11:55.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:54 vm09 bash[21220]: cluster 2026-03-07T07:11:53.939812+0000 mgr.vm07.yrfcuj (mgr.14201) 948 : cluster [DBG] pgmap v560: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:11:55.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:54 vm09 bash[21220]: cluster 2026-03-07T07:11:53.939812+0000 mgr.vm07.yrfcuj (mgr.14201) 948 : cluster [DBG] pgmap v560: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:11:55.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:54 vm09 bash[21220]: audit 2026-03-07T07:11:53.944089+0000 mon.vm07 (mon.0) 1180 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:11:55.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:54 vm09 bash[21220]: audit 2026-03-07T07:11:53.944089+0000 mon.vm07 (mon.0) 1180 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:11:55.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:54 vm09 bash[21220]: audit 2026-03-07T07:11:53.946941+0000 mon.vm07 (mon.0) 1181 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:11:55.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:54 vm09 bash[21220]: audit 2026-03-07T07:11:53.946941+0000 mon.vm07 (mon.0) 1181 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:11:55.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:54 vm07 bash[17031]: audit 2026-03-07T07:11:53.936087+0000 mon.vm07 (mon.0) 1178 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:11:55.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:54 vm07 bash[17031]: audit 2026-03-07T07:11:53.936087+0000 mon.vm07 (mon.0) 1178 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:11:55.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:54 vm07 bash[17031]: audit 2026-03-07T07:11:53.937152+0000 mon.vm07 (mon.0) 1179 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:11:55.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:54 vm07 bash[17031]: audit 2026-03-07T07:11:53.937152+0000 mon.vm07 (mon.0) 1179 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:11:55.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:54 vm07 bash[17031]: cluster 2026-03-07T07:11:53.938834+0000 mgr.vm07.yrfcuj (mgr.14201) 947 : cluster [DBG] pgmap v559: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:11:55.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:54 vm07 bash[17031]: cluster 2026-03-07T07:11:53.938834+0000 mgr.vm07.yrfcuj (mgr.14201) 947 : cluster [DBG] pgmap v559: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:11:55.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:54 vm07 bash[17031]: cluster 2026-03-07T07:11:53.939812+0000 mgr.vm07.yrfcuj (mgr.14201) 948 : cluster [DBG] pgmap v560: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:11:55.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:54 vm07 bash[17031]: cluster 2026-03-07T07:11:53.939812+0000 mgr.vm07.yrfcuj (mgr.14201) 948 : cluster [DBG] pgmap v560: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:11:55.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:54 vm07 bash[17031]: audit 2026-03-07T07:11:53.944089+0000 mon.vm07 (mon.0) 1180 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:11:55.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:54 vm07 bash[17031]: audit 2026-03-07T07:11:53.944089+0000 mon.vm07 (mon.0) 1180 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:11:55.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:54 vm07 bash[17031]: audit 2026-03-07T07:11:53.946941+0000 mon.vm07 (mon.0) 1181 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:11:55.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:54 vm07 bash[17031]: audit 2026-03-07T07:11:53.946941+0000 mon.vm07 (mon.0) 1181 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:11:56.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:55 vm09 bash[21220]: audit 2026-03-07T07:11:55.143344+0000 mon.vm07 (mon.0) 1182 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:11:56.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:55 vm09 bash[21220]: audit 2026-03-07T07:11:55.143344+0000 mon.vm07 (mon.0) 1182 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:11:56.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:55 vm07 bash[17031]: audit 2026-03-07T07:11:55.143344+0000 mon.vm07 (mon.0) 1182 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:11:56.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:55 vm07 bash[17031]: audit 2026-03-07T07:11:55.143344+0000 mon.vm07 (mon.0) 1182 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:11:57.053 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:56 vm07 bash[17031]: cluster 2026-03-07T07:11:55.940185+0000 mgr.vm07.yrfcuj (mgr.14201) 949 : cluster [DBG] pgmap v561: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:11:57.053 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:56 vm07 bash[17031]: cluster 2026-03-07T07:11:55.940185+0000 mgr.vm07.yrfcuj (mgr.14201) 949 : cluster [DBG] pgmap v561: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:11:57.091 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:11:57.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:56 vm09 bash[21220]: cluster 2026-03-07T07:11:55.940185+0000 mgr.vm07.yrfcuj (mgr.14201) 949 : cluster [DBG] pgmap v561: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:11:57.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:56 vm09 bash[21220]: cluster 2026-03-07T07:11:55.940185+0000 mgr.vm07.yrfcuj (mgr.14201) 949 : cluster [DBG] pgmap v561: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:11:57.268 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:11:57.268 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (8m) 7m ago 14m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:11:57.268 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (7m) 7m ago 14m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:11:57.268 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 2m ago 14m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:11:57.268 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 2m ago 14m - - 2026-03-07T08:11:57.476 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:11:57.476 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:11:57.476 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:11:58.101 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:57 vm09 bash[21220]: audit 2026-03-07T07:11:57.055923+0000 mgr.vm07.yrfcuj (mgr.14201) 950 : audit [DBG] from='client.16490 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:58.101 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:57 vm09 bash[21220]: audit 2026-03-07T07:11:57.055923+0000 mgr.vm07.yrfcuj (mgr.14201) 950 : audit [DBG] from='client.16490 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:58.101 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:57 vm09 bash[21220]: audit 2026-03-07T07:11:57.269697+0000 mgr.vm07.yrfcuj (mgr.14201) 951 : audit [DBG] from='client.16494 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:58.101 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:57 vm09 bash[21220]: audit 2026-03-07T07:11:57.269697+0000 mgr.vm07.yrfcuj (mgr.14201) 951 : audit [DBG] from='client.16494 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:58.101 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:57 vm09 bash[21220]: audit 2026-03-07T07:11:57.475038+0000 mon.vm09 (mon.1) 51 : audit [DBG] from='client.? 192.168.123.107:0/701096550' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:58.101 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:57 vm09 bash[21220]: audit 2026-03-07T07:11:57.475038+0000 mon.vm09 (mon.1) 51 : audit [DBG] from='client.? 192.168.123.107:0/701096550' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:58.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:57 vm07 bash[17031]: audit 2026-03-07T07:11:57.055923+0000 mgr.vm07.yrfcuj (mgr.14201) 950 : audit [DBG] from='client.16490 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:58.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:57 vm07 bash[17031]: audit 2026-03-07T07:11:57.055923+0000 mgr.vm07.yrfcuj (mgr.14201) 950 : audit [DBG] from='client.16490 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:58.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:57 vm07 bash[17031]: audit 2026-03-07T07:11:57.269697+0000 mgr.vm07.yrfcuj (mgr.14201) 951 : audit [DBG] from='client.16494 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:58.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:57 vm07 bash[17031]: audit 2026-03-07T07:11:57.269697+0000 mgr.vm07.yrfcuj (mgr.14201) 951 : audit [DBG] from='client.16494 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:11:58.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:57 vm07 bash[17031]: audit 2026-03-07T07:11:57.475038+0000 mon.vm09 (mon.1) 51 : audit [DBG] from='client.? 192.168.123.107:0/701096550' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:58.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:57 vm07 bash[17031]: audit 2026-03-07T07:11:57.475038+0000 mon.vm09 (mon.1) 51 : audit [DBG] from='client.? 192.168.123.107:0/701096550' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:11:59.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:58 vm09 bash[21220]: cluster 2026-03-07T07:11:57.940665+0000 mgr.vm07.yrfcuj (mgr.14201) 952 : cluster [DBG] pgmap v562: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:11:59.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:11:58 vm09 bash[21220]: cluster 2026-03-07T07:11:57.940665+0000 mgr.vm07.yrfcuj (mgr.14201) 952 : cluster [DBG] pgmap v562: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:11:59.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:58 vm07 bash[17031]: cluster 2026-03-07T07:11:57.940665+0000 mgr.vm07.yrfcuj (mgr.14201) 952 : cluster [DBG] pgmap v562: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:11:59.148 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:11:58 vm07 bash[17031]: cluster 2026-03-07T07:11:57.940665+0000 mgr.vm07.yrfcuj (mgr.14201) 952 : cluster [DBG] pgmap v562: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:12:01.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:00 vm09 bash[21220]: cluster 2026-03-07T07:11:59.941053+0000 mgr.vm07.yrfcuj (mgr.14201) 953 : cluster [DBG] pgmap v563: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:12:01.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:00 vm09 bash[21220]: cluster 2026-03-07T07:11:59.941053+0000 mgr.vm07.yrfcuj (mgr.14201) 953 : cluster [DBG] pgmap v563: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:12:01.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:00 vm07 bash[17031]: cluster 2026-03-07T07:11:59.941053+0000 mgr.vm07.yrfcuj (mgr.14201) 953 : cluster [DBG] pgmap v563: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:12:01.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:00 vm07 bash[17031]: cluster 2026-03-07T07:11:59.941053+0000 mgr.vm07.yrfcuj (mgr.14201) 953 : cluster [DBG] pgmap v563: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:12:02.675 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:12:02.831 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:12:02.831 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (8m) 7m ago 14m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:12:02.831 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (7m) 7m ago 14m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:12:02.831 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 2m ago 14m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:12:02.831 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 2m ago 14m - - 2026-03-07T08:12:03.032 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:12:03.032 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:12:03.032 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:12:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:02 vm09 bash[21220]: cluster 2026-03-07T07:12:01.941486+0000 mgr.vm07.yrfcuj (mgr.14201) 954 : cluster [DBG] pgmap v564: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 244 B/s rd, 488 B/s wr, 0 op/s 2026-03-07T08:12:03.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:02 vm09 bash[21220]: cluster 2026-03-07T07:12:01.941486+0000 mgr.vm07.yrfcuj (mgr.14201) 954 : cluster [DBG] pgmap v564: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 244 B/s rd, 488 B/s wr, 0 op/s 2026-03-07T08:12:03.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:03 vm07 bash[17031]: cluster 2026-03-07T07:12:01.941486+0000 mgr.vm07.yrfcuj (mgr.14201) 954 : cluster [DBG] pgmap v564: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 244 B/s rd, 488 B/s wr, 0 op/s 2026-03-07T08:12:03.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:03 vm07 bash[17031]: cluster 2026-03-07T07:12:01.941486+0000 mgr.vm07.yrfcuj (mgr.14201) 954 : cluster [DBG] pgmap v564: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 244 B/s rd, 488 B/s wr, 0 op/s 2026-03-07T08:12:04.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:04 vm09 bash[21220]: audit 2026-03-07T07:12:02.662166+0000 mgr.vm07.yrfcuj (mgr.14201) 955 : audit [DBG] from='client.16502 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:04.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:04 vm09 bash[21220]: audit 2026-03-07T07:12:02.662166+0000 mgr.vm07.yrfcuj (mgr.14201) 955 : audit [DBG] from='client.16502 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:04.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:04 vm09 bash[21220]: audit 2026-03-07T07:12:02.833223+0000 mgr.vm07.yrfcuj (mgr.14201) 956 : audit [DBG] from='client.25647 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:04.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:04 vm09 bash[21220]: audit 2026-03-07T07:12:02.833223+0000 mgr.vm07.yrfcuj (mgr.14201) 956 : audit [DBG] from='client.25647 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:04.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:04 vm09 bash[21220]: audit 2026-03-07T07:12:03.036787+0000 mon.vm07 (mon.0) 1183 : audit [DBG] from='client.? 192.168.123.107:0/3985748013' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:04.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:04 vm09 bash[21220]: audit 2026-03-07T07:12:03.036787+0000 mon.vm07 (mon.0) 1183 : audit [DBG] from='client.? 192.168.123.107:0/3985748013' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:04.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:04 vm07 bash[17031]: audit 2026-03-07T07:12:02.662166+0000 mgr.vm07.yrfcuj (mgr.14201) 955 : audit [DBG] from='client.16502 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:04.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:04 vm07 bash[17031]: audit 2026-03-07T07:12:02.662166+0000 mgr.vm07.yrfcuj (mgr.14201) 955 : audit [DBG] from='client.16502 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:04.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:04 vm07 bash[17031]: audit 2026-03-07T07:12:02.833223+0000 mgr.vm07.yrfcuj (mgr.14201) 956 : audit [DBG] from='client.25647 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:04.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:04 vm07 bash[17031]: audit 2026-03-07T07:12:02.833223+0000 mgr.vm07.yrfcuj (mgr.14201) 956 : audit [DBG] from='client.25647 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:04.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:04 vm07 bash[17031]: audit 2026-03-07T07:12:03.036787+0000 mon.vm07 (mon.0) 1183 : audit [DBG] from='client.? 192.168.123.107:0/3985748013' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:04.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:04 vm07 bash[17031]: audit 2026-03-07T07:12:03.036787+0000 mon.vm07 (mon.0) 1183 : audit [DBG] from='client.? 192.168.123.107:0/3985748013' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:05.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:05 vm09 bash[21220]: cluster 2026-03-07T07:12:03.941897+0000 mgr.vm07.yrfcuj (mgr.14201) 957 : cluster [DBG] pgmap v565: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 204 B/s rd, 409 B/s wr, 0 op/s 2026-03-07T08:12:05.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:05 vm09 bash[21220]: cluster 2026-03-07T07:12:03.941897+0000 mgr.vm07.yrfcuj (mgr.14201) 957 : cluster [DBG] pgmap v565: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 204 B/s rd, 409 B/s wr, 0 op/s 2026-03-07T08:12:05.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:05 vm07 bash[17031]: cluster 2026-03-07T07:12:03.941897+0000 mgr.vm07.yrfcuj (mgr.14201) 957 : cluster [DBG] pgmap v565: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 204 B/s rd, 409 B/s wr, 0 op/s 2026-03-07T08:12:05.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:05 vm07 bash[17031]: cluster 2026-03-07T07:12:03.941897+0000 mgr.vm07.yrfcuj (mgr.14201) 957 : cluster [DBG] pgmap v565: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 204 B/s rd, 409 B/s wr, 0 op/s 2026-03-07T08:12:07.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:07 vm09 bash[21220]: cluster 2026-03-07T07:12:05.942276+0000 mgr.vm07.yrfcuj (mgr.14201) 958 : cluster [DBG] pgmap v566: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:07.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:07 vm09 bash[21220]: cluster 2026-03-07T07:12:05.942276+0000 mgr.vm07.yrfcuj (mgr.14201) 958 : cluster [DBG] pgmap v566: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:07.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:07 vm07 bash[17031]: cluster 2026-03-07T07:12:05.942276+0000 mgr.vm07.yrfcuj (mgr.14201) 958 : cluster [DBG] pgmap v566: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:07.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:07 vm07 bash[17031]: cluster 2026-03-07T07:12:05.942276+0000 mgr.vm07.yrfcuj (mgr.14201) 958 : cluster [DBG] pgmap v566: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:08.220 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:12:08.388 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:12:08.388 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (8m) 7m ago 14m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:12:08.388 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (8m) 7m ago 14m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:12:08.388 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 2m ago 14m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:12:08.388 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 2m ago 14m - - 2026-03-07T08:12:08.591 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:12:08.591 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:12:08.591 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:12:09.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:09 vm07 bash[17031]: cluster 2026-03-07T07:12:07.942773+0000 mgr.vm07.yrfcuj (mgr.14201) 959 : cluster [DBG] pgmap v567: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:09.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:09 vm07 bash[17031]: cluster 2026-03-07T07:12:07.942773+0000 mgr.vm07.yrfcuj (mgr.14201) 959 : cluster [DBG] pgmap v567: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:09.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:09 vm07 bash[17031]: audit 2026-03-07T07:12:08.206577+0000 mgr.vm07.yrfcuj (mgr.14201) 960 : audit [DBG] from='client.16514 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:09.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:09 vm07 bash[17031]: audit 2026-03-07T07:12:08.206577+0000 mgr.vm07.yrfcuj (mgr.14201) 960 : audit [DBG] from='client.16514 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:09.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:09 vm07 bash[17031]: audit 2026-03-07T07:12:08.389817+0000 mgr.vm07.yrfcuj (mgr.14201) 961 : audit [DBG] from='client.16518 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:09.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:09 vm07 bash[17031]: audit 2026-03-07T07:12:08.389817+0000 mgr.vm07.yrfcuj (mgr.14201) 961 : audit [DBG] from='client.16518 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:09.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:09 vm07 bash[17031]: audit 2026-03-07T07:12:08.595326+0000 mon.vm07 (mon.0) 1184 : audit [DBG] from='client.? 192.168.123.107:0/1377463926' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:09.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:09 vm07 bash[17031]: audit 2026-03-07T07:12:08.595326+0000 mon.vm07 (mon.0) 1184 : audit [DBG] from='client.? 192.168.123.107:0/1377463926' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:09.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:09 vm09 bash[21220]: cluster 2026-03-07T07:12:07.942773+0000 mgr.vm07.yrfcuj (mgr.14201) 959 : cluster [DBG] pgmap v567: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:09.616 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:09 vm09 bash[21220]: cluster 2026-03-07T07:12:07.942773+0000 mgr.vm07.yrfcuj (mgr.14201) 959 : cluster [DBG] pgmap v567: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:09.616 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:09 vm09 bash[21220]: audit 2026-03-07T07:12:08.206577+0000 mgr.vm07.yrfcuj (mgr.14201) 960 : audit [DBG] from='client.16514 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:09.616 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:09 vm09 bash[21220]: audit 2026-03-07T07:12:08.206577+0000 mgr.vm07.yrfcuj (mgr.14201) 960 : audit [DBG] from='client.16514 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:09.616 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:09 vm09 bash[21220]: audit 2026-03-07T07:12:08.389817+0000 mgr.vm07.yrfcuj (mgr.14201) 961 : audit [DBG] from='client.16518 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:09.616 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:09 vm09 bash[21220]: audit 2026-03-07T07:12:08.389817+0000 mgr.vm07.yrfcuj (mgr.14201) 961 : audit [DBG] from='client.16518 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:09.616 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:09 vm09 bash[21220]: audit 2026-03-07T07:12:08.595326+0000 mon.vm07 (mon.0) 1184 : audit [DBG] from='client.? 192.168.123.107:0/1377463926' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:09.616 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:09 vm09 bash[21220]: audit 2026-03-07T07:12:08.595326+0000 mon.vm07 (mon.0) 1184 : audit [DBG] from='client.? 192.168.123.107:0/1377463926' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:11.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:11 vm07 bash[17031]: cluster 2026-03-07T07:12:09.943242+0000 mgr.vm07.yrfcuj (mgr.14201) 962 : cluster [DBG] pgmap v568: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:11.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:11 vm07 bash[17031]: cluster 2026-03-07T07:12:09.943242+0000 mgr.vm07.yrfcuj (mgr.14201) 962 : cluster [DBG] pgmap v568: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:11.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:11 vm07 bash[17031]: audit 2026-03-07T07:12:10.143398+0000 mon.vm07 (mon.0) 1185 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:12:11.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:11 vm07 bash[17031]: audit 2026-03-07T07:12:10.143398+0000 mon.vm07 (mon.0) 1185 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:12:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:11 vm09 bash[21220]: cluster 2026-03-07T07:12:09.943242+0000 mgr.vm07.yrfcuj (mgr.14201) 962 : cluster [DBG] pgmap v568: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:11 vm09 bash[21220]: cluster 2026-03-07T07:12:09.943242+0000 mgr.vm07.yrfcuj (mgr.14201) 962 : cluster [DBG] pgmap v568: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:11 vm09 bash[21220]: audit 2026-03-07T07:12:10.143398+0000 mon.vm07 (mon.0) 1185 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:12:11.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:11 vm09 bash[21220]: audit 2026-03-07T07:12:10.143398+0000 mon.vm07 (mon.0) 1185 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:12:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:13 vm09 bash[21220]: cluster 2026-03-07T07:12:11.943691+0000 mgr.vm07.yrfcuj (mgr.14201) 963 : cluster [DBG] pgmap v569: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:13.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:13 vm09 bash[21220]: cluster 2026-03-07T07:12:11.943691+0000 mgr.vm07.yrfcuj (mgr.14201) 963 : cluster [DBG] pgmap v569: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:13.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:13 vm07 bash[17031]: cluster 2026-03-07T07:12:11.943691+0000 mgr.vm07.yrfcuj (mgr.14201) 963 : cluster [DBG] pgmap v569: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:13.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:13 vm07 bash[17031]: cluster 2026-03-07T07:12:11.943691+0000 mgr.vm07.yrfcuj (mgr.14201) 963 : cluster [DBG] pgmap v569: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:13.785 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:12:13.947 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:12:13.947 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (8m) 8m ago 14m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:12:13.947 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (8m) 8m ago 14m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:12:13.947 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 2m ago 14m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:12:13.947 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 2m ago 14m - - 2026-03-07T08:12:14.153 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:12:14.153 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:12:14.153 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:12:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:14 vm09 bash[21220]: audit 2026-03-07T07:12:14.157212+0000 mon.vm07 (mon.0) 1186 : audit [DBG] from='client.? 192.168.123.107:0/983140625' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:14.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:14 vm09 bash[21220]: audit 2026-03-07T07:12:14.157212+0000 mon.vm07 (mon.0) 1186 : audit [DBG] from='client.? 192.168.123.107:0/983140625' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:14.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:14 vm07 bash[17031]: audit 2026-03-07T07:12:14.157212+0000 mon.vm07 (mon.0) 1186 : audit [DBG] from='client.? 192.168.123.107:0/983140625' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:14.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:14 vm07 bash[17031]: audit 2026-03-07T07:12:14.157212+0000 mon.vm07 (mon.0) 1186 : audit [DBG] from='client.? 192.168.123.107:0/983140625' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:15.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:15 vm07 bash[17031]: audit 2026-03-07T07:12:13.772911+0000 mgr.vm07.yrfcuj (mgr.14201) 964 : audit [DBG] from='client.16526 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:15.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:15 vm07 bash[17031]: audit 2026-03-07T07:12:13.772911+0000 mgr.vm07.yrfcuj (mgr.14201) 964 : audit [DBG] from='client.16526 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:15.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:15 vm07 bash[17031]: cluster 2026-03-07T07:12:13.944111+0000 mgr.vm07.yrfcuj (mgr.14201) 965 : cluster [DBG] pgmap v570: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:15.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:15 vm07 bash[17031]: cluster 2026-03-07T07:12:13.944111+0000 mgr.vm07.yrfcuj (mgr.14201) 965 : cluster [DBG] pgmap v570: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:15.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:15 vm07 bash[17031]: audit 2026-03-07T07:12:13.948726+0000 mgr.vm07.yrfcuj (mgr.14201) 966 : audit [DBG] from='client.16530 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:15.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:15 vm07 bash[17031]: audit 2026-03-07T07:12:13.948726+0000 mgr.vm07.yrfcuj (mgr.14201) 966 : audit [DBG] from='client.16530 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:15.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:15 vm09 bash[21220]: audit 2026-03-07T07:12:13.772911+0000 mgr.vm07.yrfcuj (mgr.14201) 964 : audit [DBG] from='client.16526 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:15.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:15 vm09 bash[21220]: audit 2026-03-07T07:12:13.772911+0000 mgr.vm07.yrfcuj (mgr.14201) 964 : audit [DBG] from='client.16526 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:15.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:15 vm09 bash[21220]: cluster 2026-03-07T07:12:13.944111+0000 mgr.vm07.yrfcuj (mgr.14201) 965 : cluster [DBG] pgmap v570: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:15.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:15 vm09 bash[21220]: cluster 2026-03-07T07:12:13.944111+0000 mgr.vm07.yrfcuj (mgr.14201) 965 : cluster [DBG] pgmap v570: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:15.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:15 vm09 bash[21220]: audit 2026-03-07T07:12:13.948726+0000 mgr.vm07.yrfcuj (mgr.14201) 966 : audit [DBG] from='client.16530 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:15.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:15 vm09 bash[21220]: audit 2026-03-07T07:12:13.948726+0000 mgr.vm07.yrfcuj (mgr.14201) 966 : audit [DBG] from='client.16530 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:17.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:17 vm09 bash[21220]: cluster 2026-03-07T07:12:15.944637+0000 mgr.vm07.yrfcuj (mgr.14201) 967 : cluster [DBG] pgmap v571: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:17.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:17 vm09 bash[21220]: cluster 2026-03-07T07:12:15.944637+0000 mgr.vm07.yrfcuj (mgr.14201) 967 : cluster [DBG] pgmap v571: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:17.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:17 vm07 bash[17031]: cluster 2026-03-07T07:12:15.944637+0000 mgr.vm07.yrfcuj (mgr.14201) 967 : cluster [DBG] pgmap v571: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:17.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:17 vm07 bash[17031]: cluster 2026-03-07T07:12:15.944637+0000 mgr.vm07.yrfcuj (mgr.14201) 967 : cluster [DBG] pgmap v571: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:19.355 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:12:19.519 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:12:19.519 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (8m) 8m ago 14m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:12:19.519 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (8m) 8m ago 14m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:12:19.519 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (2m) 2m ago 14m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:12:19.519 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 2m ago 14m - - 2026-03-07T08:12:19.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:19 vm09 bash[21220]: cluster 2026-03-07T07:12:17.945307+0000 mgr.vm07.yrfcuj (mgr.14201) 968 : cluster [DBG] pgmap v572: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:19.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:19 vm09 bash[21220]: cluster 2026-03-07T07:12:17.945307+0000 mgr.vm07.yrfcuj (mgr.14201) 968 : cluster [DBG] pgmap v572: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:19.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:19 vm07 bash[17031]: cluster 2026-03-07T07:12:17.945307+0000 mgr.vm07.yrfcuj (mgr.14201) 968 : cluster [DBG] pgmap v572: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:19.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:19 vm07 bash[17031]: cluster 2026-03-07T07:12:17.945307+0000 mgr.vm07.yrfcuj (mgr.14201) 968 : cluster [DBG] pgmap v572: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:19.726 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:12:19.726 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:12:19.726 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:12:20.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:20 vm07 bash[17031]: audit 2026-03-07T07:12:19.344240+0000 mgr.vm07.yrfcuj (mgr.14201) 969 : audit [DBG] from='client.16538 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:20.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:20 vm07 bash[17031]: audit 2026-03-07T07:12:19.344240+0000 mgr.vm07.yrfcuj (mgr.14201) 969 : audit [DBG] from='client.16538 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:20.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:20 vm07 bash[17031]: audit 2026-03-07T07:12:19.521355+0000 mgr.vm07.yrfcuj (mgr.14201) 970 : audit [DBG] from='client.16542 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:20.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:20 vm07 bash[17031]: audit 2026-03-07T07:12:19.521355+0000 mgr.vm07.yrfcuj (mgr.14201) 970 : audit [DBG] from='client.16542 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:20.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:20 vm07 bash[17031]: audit 2026-03-07T07:12:19.730429+0000 mon.vm07 (mon.0) 1187 : audit [DBG] from='client.? 192.168.123.107:0/3578195752' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:20.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:20 vm07 bash[17031]: audit 2026-03-07T07:12:19.730429+0000 mon.vm07 (mon.0) 1187 : audit [DBG] from='client.? 192.168.123.107:0/3578195752' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:20 vm09 bash[21220]: audit 2026-03-07T07:12:19.344240+0000 mgr.vm07.yrfcuj (mgr.14201) 969 : audit [DBG] from='client.16538 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:20 vm09 bash[21220]: audit 2026-03-07T07:12:19.344240+0000 mgr.vm07.yrfcuj (mgr.14201) 969 : audit [DBG] from='client.16538 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:20 vm09 bash[21220]: audit 2026-03-07T07:12:19.521355+0000 mgr.vm07.yrfcuj (mgr.14201) 970 : audit [DBG] from='client.16542 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:20 vm09 bash[21220]: audit 2026-03-07T07:12:19.521355+0000 mgr.vm07.yrfcuj (mgr.14201) 970 : audit [DBG] from='client.16542 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:20 vm09 bash[21220]: audit 2026-03-07T07:12:19.730429+0000 mon.vm07 (mon.0) 1187 : audit [DBG] from='client.? 192.168.123.107:0/3578195752' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:20.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:20 vm09 bash[21220]: audit 2026-03-07T07:12:19.730429+0000 mon.vm07 (mon.0) 1187 : audit [DBG] from='client.? 192.168.123.107:0/3578195752' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:21.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:21 vm07 bash[17031]: cluster 2026-03-07T07:12:19.945707+0000 mgr.vm07.yrfcuj (mgr.14201) 971 : cluster [DBG] pgmap v573: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:21.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:21 vm07 bash[17031]: cluster 2026-03-07T07:12:19.945707+0000 mgr.vm07.yrfcuj (mgr.14201) 971 : cluster [DBG] pgmap v573: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:21.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:21 vm09 bash[21220]: cluster 2026-03-07T07:12:19.945707+0000 mgr.vm07.yrfcuj (mgr.14201) 971 : cluster [DBG] pgmap v573: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:21.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:21 vm09 bash[21220]: cluster 2026-03-07T07:12:19.945707+0000 mgr.vm07.yrfcuj (mgr.14201) 971 : cluster [DBG] pgmap v573: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:23.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:23 vm07 bash[17031]: cluster 2026-03-07T07:12:21.946122+0000 mgr.vm07.yrfcuj (mgr.14201) 972 : cluster [DBG] pgmap v574: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:23.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:23 vm07 bash[17031]: cluster 2026-03-07T07:12:21.946122+0000 mgr.vm07.yrfcuj (mgr.14201) 972 : cluster [DBG] pgmap v574: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:23.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:23 vm09 bash[21220]: cluster 2026-03-07T07:12:21.946122+0000 mgr.vm07.yrfcuj (mgr.14201) 972 : cluster [DBG] pgmap v574: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:23.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:23 vm09 bash[21220]: cluster 2026-03-07T07:12:21.946122+0000 mgr.vm07.yrfcuj (mgr.14201) 972 : cluster [DBG] pgmap v574: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:24.906 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:12:25.067 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:12:25.068 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (8m) 8m ago 14m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:12:25.068 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (8m) 8m ago 14m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:12:25.068 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 2m ago 14m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:12:25.068 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 2m ago 14m - - 2026-03-07T08:12:25.288 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:12:25.289 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:12:25.289 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:12:25.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:25 vm07 bash[17031]: cluster 2026-03-07T07:12:23.946570+0000 mgr.vm07.yrfcuj (mgr.14201) 973 : cluster [DBG] pgmap v575: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:25.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:25 vm07 bash[17031]: cluster 2026-03-07T07:12:23.946570+0000 mgr.vm07.yrfcuj (mgr.14201) 973 : cluster [DBG] pgmap v575: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:25.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:25 vm07 bash[17031]: audit 2026-03-07T07:12:25.143661+0000 mon.vm07 (mon.0) 1188 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:12:25.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:25 vm07 bash[17031]: audit 2026-03-07T07:12:25.143661+0000 mon.vm07 (mon.0) 1188 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:12:25.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:25 vm07 bash[17031]: audit 2026-03-07T07:12:25.293260+0000 mon.vm07 (mon.0) 1189 : audit [DBG] from='client.? 192.168.123.107:0/424687557' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:25.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:25 vm07 bash[17031]: audit 2026-03-07T07:12:25.293260+0000 mon.vm07 (mon.0) 1189 : audit [DBG] from='client.? 192.168.123.107:0/424687557' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:25 vm09 bash[21220]: cluster 2026-03-07T07:12:23.946570+0000 mgr.vm07.yrfcuj (mgr.14201) 973 : cluster [DBG] pgmap v575: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:25 vm09 bash[21220]: cluster 2026-03-07T07:12:23.946570+0000 mgr.vm07.yrfcuj (mgr.14201) 973 : cluster [DBG] pgmap v575: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:25 vm09 bash[21220]: audit 2026-03-07T07:12:25.143661+0000 mon.vm07 (mon.0) 1188 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:12:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:25 vm09 bash[21220]: audit 2026-03-07T07:12:25.143661+0000 mon.vm07 (mon.0) 1188 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:12:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:25 vm09 bash[21220]: audit 2026-03-07T07:12:25.293260+0000 mon.vm07 (mon.0) 1189 : audit [DBG] from='client.? 192.168.123.107:0/424687557' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:25 vm09 bash[21220]: audit 2026-03-07T07:12:25.293260+0000 mon.vm07 (mon.0) 1189 : audit [DBG] from='client.? 192.168.123.107:0/424687557' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:26.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:26 vm09 bash[21220]: audit 2026-03-07T07:12:24.893828+0000 mgr.vm07.yrfcuj (mgr.14201) 974 : audit [DBG] from='client.16550 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:26.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:26 vm09 bash[21220]: audit 2026-03-07T07:12:24.893828+0000 mgr.vm07.yrfcuj (mgr.14201) 974 : audit [DBG] from='client.16550 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:26.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:26 vm09 bash[21220]: audit 2026-03-07T07:12:25.068629+0000 mgr.vm07.yrfcuj (mgr.14201) 975 : audit [DBG] from='client.16554 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:26.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:26 vm09 bash[21220]: audit 2026-03-07T07:12:25.068629+0000 mgr.vm07.yrfcuj (mgr.14201) 975 : audit [DBG] from='client.16554 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:26.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:26 vm07 bash[17031]: audit 2026-03-07T07:12:24.893828+0000 mgr.vm07.yrfcuj (mgr.14201) 974 : audit [DBG] from='client.16550 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:26.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:26 vm07 bash[17031]: audit 2026-03-07T07:12:24.893828+0000 mgr.vm07.yrfcuj (mgr.14201) 974 : audit [DBG] from='client.16550 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:26.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:26 vm07 bash[17031]: audit 2026-03-07T07:12:25.068629+0000 mgr.vm07.yrfcuj (mgr.14201) 975 : audit [DBG] from='client.16554 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:26.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:26 vm07 bash[17031]: audit 2026-03-07T07:12:25.068629+0000 mgr.vm07.yrfcuj (mgr.14201) 975 : audit [DBG] from='client.16554 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:27.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:27 vm09 bash[21220]: cluster 2026-03-07T07:12:25.947094+0000 mgr.vm07.yrfcuj (mgr.14201) 976 : cluster [DBG] pgmap v576: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:27.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:27 vm09 bash[21220]: cluster 2026-03-07T07:12:25.947094+0000 mgr.vm07.yrfcuj (mgr.14201) 976 : cluster [DBG] pgmap v576: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:27.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:27 vm07 bash[17031]: cluster 2026-03-07T07:12:25.947094+0000 mgr.vm07.yrfcuj (mgr.14201) 976 : cluster [DBG] pgmap v576: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:27.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:27 vm07 bash[17031]: cluster 2026-03-07T07:12:25.947094+0000 mgr.vm07.yrfcuj (mgr.14201) 976 : cluster [DBG] pgmap v576: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:29.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:29 vm09 bash[21220]: cluster 2026-03-07T07:12:27.947743+0000 mgr.vm07.yrfcuj (mgr.14201) 977 : cluster [DBG] pgmap v577: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:29.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:29 vm09 bash[21220]: cluster 2026-03-07T07:12:27.947743+0000 mgr.vm07.yrfcuj (mgr.14201) 977 : cluster [DBG] pgmap v577: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:29.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:29 vm07 bash[17031]: cluster 2026-03-07T07:12:27.947743+0000 mgr.vm07.yrfcuj (mgr.14201) 977 : cluster [DBG] pgmap v577: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:29.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:29 vm07 bash[17031]: cluster 2026-03-07T07:12:27.947743+0000 mgr.vm07.yrfcuj (mgr.14201) 977 : cluster [DBG] pgmap v577: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:30.474 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:12:30.634 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:12:30.634 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (9m) 8m ago 14m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:12:30.634 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (8m) 8m ago 14m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:12:30.634 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 2m ago 14m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:12:30.634 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 2m ago 14m - - 2026-03-07T08:12:30.823 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:12:30.823 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:12:30.823 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:12:31.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:31 vm09 bash[21220]: cluster 2026-03-07T07:12:29.948150+0000 mgr.vm07.yrfcuj (mgr.14201) 978 : cluster [DBG] pgmap v578: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:31.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:31 vm09 bash[21220]: cluster 2026-03-07T07:12:29.948150+0000 mgr.vm07.yrfcuj (mgr.14201) 978 : cluster [DBG] pgmap v578: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:31.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:31 vm09 bash[21220]: audit 2026-03-07T07:12:30.462289+0000 mgr.vm07.yrfcuj (mgr.14201) 979 : audit [DBG] from='client.16562 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:31.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:31 vm09 bash[21220]: audit 2026-03-07T07:12:30.462289+0000 mgr.vm07.yrfcuj (mgr.14201) 979 : audit [DBG] from='client.16562 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:31.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:31 vm09 bash[21220]: audit 2026-03-07T07:12:30.827416+0000 mon.vm07 (mon.0) 1190 : audit [DBG] from='client.? 192.168.123.107:0/3266662148' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:31.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:31 vm09 bash[21220]: audit 2026-03-07T07:12:30.827416+0000 mon.vm07 (mon.0) 1190 : audit [DBG] from='client.? 192.168.123.107:0/3266662148' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:31.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:31 vm07 bash[17031]: cluster 2026-03-07T07:12:29.948150+0000 mgr.vm07.yrfcuj (mgr.14201) 978 : cluster [DBG] pgmap v578: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:31.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:31 vm07 bash[17031]: cluster 2026-03-07T07:12:29.948150+0000 mgr.vm07.yrfcuj (mgr.14201) 978 : cluster [DBG] pgmap v578: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:31.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:31 vm07 bash[17031]: audit 2026-03-07T07:12:30.462289+0000 mgr.vm07.yrfcuj (mgr.14201) 979 : audit [DBG] from='client.16562 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:31.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:31 vm07 bash[17031]: audit 2026-03-07T07:12:30.462289+0000 mgr.vm07.yrfcuj (mgr.14201) 979 : audit [DBG] from='client.16562 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:31.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:31 vm07 bash[17031]: audit 2026-03-07T07:12:30.827416+0000 mon.vm07 (mon.0) 1190 : audit [DBG] from='client.? 192.168.123.107:0/3266662148' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:31.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:31 vm07 bash[17031]: audit 2026-03-07T07:12:30.827416+0000 mon.vm07 (mon.0) 1190 : audit [DBG] from='client.? 192.168.123.107:0/3266662148' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:32.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:32 vm09 bash[21220]: audit 2026-03-07T07:12:30.634918+0000 mgr.vm07.yrfcuj (mgr.14201) 980 : audit [DBG] from='client.16566 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:32.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:32 vm09 bash[21220]: audit 2026-03-07T07:12:30.634918+0000 mgr.vm07.yrfcuj (mgr.14201) 980 : audit [DBG] from='client.16566 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:32.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:32 vm07 bash[17031]: audit 2026-03-07T07:12:30.634918+0000 mgr.vm07.yrfcuj (mgr.14201) 980 : audit [DBG] from='client.16566 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:32.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:32 vm07 bash[17031]: audit 2026-03-07T07:12:30.634918+0000 mgr.vm07.yrfcuj (mgr.14201) 980 : audit [DBG] from='client.16566 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:33.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:33 vm09 bash[21220]: cluster 2026-03-07T07:12:31.948578+0000 mgr.vm07.yrfcuj (mgr.14201) 981 : cluster [DBG] pgmap v579: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:33.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:33 vm09 bash[21220]: cluster 2026-03-07T07:12:31.948578+0000 mgr.vm07.yrfcuj (mgr.14201) 981 : cluster [DBG] pgmap v579: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:33.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:33 vm07 bash[17031]: cluster 2026-03-07T07:12:31.948578+0000 mgr.vm07.yrfcuj (mgr.14201) 981 : cluster [DBG] pgmap v579: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:33.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:33 vm07 bash[17031]: cluster 2026-03-07T07:12:31.948578+0000 mgr.vm07.yrfcuj (mgr.14201) 981 : cluster [DBG] pgmap v579: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:35.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:35 vm09 bash[21220]: cluster 2026-03-07T07:12:33.948995+0000 mgr.vm07.yrfcuj (mgr.14201) 982 : cluster [DBG] pgmap v580: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:35.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:35 vm09 bash[21220]: cluster 2026-03-07T07:12:33.948995+0000 mgr.vm07.yrfcuj (mgr.14201) 982 : cluster [DBG] pgmap v580: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:35.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:35 vm07 bash[17031]: cluster 2026-03-07T07:12:33.948995+0000 mgr.vm07.yrfcuj (mgr.14201) 982 : cluster [DBG] pgmap v580: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:35.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:35 vm07 bash[17031]: cluster 2026-03-07T07:12:33.948995+0000 mgr.vm07.yrfcuj (mgr.14201) 982 : cluster [DBG] pgmap v580: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:36.008 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:12:36.168 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:12:36.168 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (9m) 8m ago 14m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:12:36.168 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (8m) 8m ago 14m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:12:36.168 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 2m ago 14m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:12:36.168 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 2m ago 14m - - 2026-03-07T08:12:36.370 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:12:36.370 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:12:36.370 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:12:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:36 vm09 bash[21220]: audit 2026-03-07T07:12:36.374716+0000 mon.vm07 (mon.0) 1191 : audit [DBG] from='client.? 192.168.123.107:0/1178907294' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:36 vm09 bash[21220]: audit 2026-03-07T07:12:36.374716+0000 mon.vm07 (mon.0) 1191 : audit [DBG] from='client.? 192.168.123.107:0/1178907294' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:36.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:36 vm07 bash[17031]: audit 2026-03-07T07:12:36.374716+0000 mon.vm07 (mon.0) 1191 : audit [DBG] from='client.? 192.168.123.107:0/1178907294' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:36.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:36 vm07 bash[17031]: audit 2026-03-07T07:12:36.374716+0000 mon.vm07 (mon.0) 1191 : audit [DBG] from='client.? 192.168.123.107:0/1178907294' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:37 vm09 bash[21220]: cluster 2026-03-07T07:12:35.949321+0000 mgr.vm07.yrfcuj (mgr.14201) 983 : cluster [DBG] pgmap v581: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:37 vm09 bash[21220]: cluster 2026-03-07T07:12:35.949321+0000 mgr.vm07.yrfcuj (mgr.14201) 983 : cluster [DBG] pgmap v581: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:37 vm09 bash[21220]: audit 2026-03-07T07:12:35.995638+0000 mgr.vm07.yrfcuj (mgr.14201) 984 : audit [DBG] from='client.16574 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:37 vm09 bash[21220]: audit 2026-03-07T07:12:35.995638+0000 mgr.vm07.yrfcuj (mgr.14201) 984 : audit [DBG] from='client.16574 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:37 vm09 bash[21220]: audit 2026-03-07T07:12:36.170041+0000 mgr.vm07.yrfcuj (mgr.14201) 985 : audit [DBG] from='client.16578 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:37 vm09 bash[21220]: audit 2026-03-07T07:12:36.170041+0000 mgr.vm07.yrfcuj (mgr.14201) 985 : audit [DBG] from='client.16578 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:37.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:37 vm07 bash[17031]: cluster 2026-03-07T07:12:35.949321+0000 mgr.vm07.yrfcuj (mgr.14201) 983 : cluster [DBG] pgmap v581: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:37.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:37 vm07 bash[17031]: cluster 2026-03-07T07:12:35.949321+0000 mgr.vm07.yrfcuj (mgr.14201) 983 : cluster [DBG] pgmap v581: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:37.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:37 vm07 bash[17031]: audit 2026-03-07T07:12:35.995638+0000 mgr.vm07.yrfcuj (mgr.14201) 984 : audit [DBG] from='client.16574 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:37.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:37 vm07 bash[17031]: audit 2026-03-07T07:12:35.995638+0000 mgr.vm07.yrfcuj (mgr.14201) 984 : audit [DBG] from='client.16574 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:37.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:37 vm07 bash[17031]: audit 2026-03-07T07:12:36.170041+0000 mgr.vm07.yrfcuj (mgr.14201) 985 : audit [DBG] from='client.16578 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:37.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:37 vm07 bash[17031]: audit 2026-03-07T07:12:36.170041+0000 mgr.vm07.yrfcuj (mgr.14201) 985 : audit [DBG] from='client.16578 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:39.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:39 vm09 bash[21220]: cluster 2026-03-07T07:12:37.949785+0000 mgr.vm07.yrfcuj (mgr.14201) 986 : cluster [DBG] pgmap v582: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:39.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:39 vm09 bash[21220]: cluster 2026-03-07T07:12:37.949785+0000 mgr.vm07.yrfcuj (mgr.14201) 986 : cluster [DBG] pgmap v582: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:39.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:39 vm07 bash[17031]: cluster 2026-03-07T07:12:37.949785+0000 mgr.vm07.yrfcuj (mgr.14201) 986 : cluster [DBG] pgmap v582: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:39.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:39 vm07 bash[17031]: cluster 2026-03-07T07:12:37.949785+0000 mgr.vm07.yrfcuj (mgr.14201) 986 : cluster [DBG] pgmap v582: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:40.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:40 vm09 bash[21220]: audit 2026-03-07T07:12:40.143868+0000 mon.vm07 (mon.0) 1192 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:12:40.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:40 vm09 bash[21220]: audit 2026-03-07T07:12:40.143868+0000 mon.vm07 (mon.0) 1192 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:12:40.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:40 vm07 bash[17031]: audit 2026-03-07T07:12:40.143868+0000 mon.vm07 (mon.0) 1192 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:12:40.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:40 vm07 bash[17031]: audit 2026-03-07T07:12:40.143868+0000 mon.vm07 (mon.0) 1192 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:12:41.549 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:12:41.705 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:12:41.705 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (9m) 8m ago 15m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:12:41.705 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (8m) 8m ago 15m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:12:41.705 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 2m ago 15m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:12:41.705 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 2m ago 15m - - 2026-03-07T08:12:41.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:41 vm09 bash[21220]: cluster 2026-03-07T07:12:39.950158+0000 mgr.vm07.yrfcuj (mgr.14201) 987 : cluster [DBG] pgmap v583: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:41.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:41 vm09 bash[21220]: cluster 2026-03-07T07:12:39.950158+0000 mgr.vm07.yrfcuj (mgr.14201) 987 : cluster [DBG] pgmap v583: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:41.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:41 vm07 bash[17031]: cluster 2026-03-07T07:12:39.950158+0000 mgr.vm07.yrfcuj (mgr.14201) 987 : cluster [DBG] pgmap v583: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:41.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:41 vm07 bash[17031]: cluster 2026-03-07T07:12:39.950158+0000 mgr.vm07.yrfcuj (mgr.14201) 987 : cluster [DBG] pgmap v583: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:41.899 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:12:41.899 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:12:41.899 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:12:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:42 vm09 bash[21220]: audit 2026-03-07T07:12:41.538902+0000 mgr.vm07.yrfcuj (mgr.14201) 988 : audit [DBG] from='client.16586 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:42 vm09 bash[21220]: audit 2026-03-07T07:12:41.538902+0000 mgr.vm07.yrfcuj (mgr.14201) 988 : audit [DBG] from='client.16586 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:42 vm09 bash[21220]: audit 2026-03-07T07:12:41.898188+0000 mon.vm09 (mon.1) 52 : audit [DBG] from='client.? 192.168.123.107:0/3872206163' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:42.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:42 vm09 bash[21220]: audit 2026-03-07T07:12:41.898188+0000 mon.vm09 (mon.1) 52 : audit [DBG] from='client.? 192.168.123.107:0/3872206163' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:42.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:42 vm07 bash[17031]: audit 2026-03-07T07:12:41.538902+0000 mgr.vm07.yrfcuj (mgr.14201) 988 : audit [DBG] from='client.16586 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:42.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:42 vm07 bash[17031]: audit 2026-03-07T07:12:41.538902+0000 mgr.vm07.yrfcuj (mgr.14201) 988 : audit [DBG] from='client.16586 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:42.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:42 vm07 bash[17031]: audit 2026-03-07T07:12:41.898188+0000 mon.vm09 (mon.1) 52 : audit [DBG] from='client.? 192.168.123.107:0/3872206163' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:42.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:42 vm07 bash[17031]: audit 2026-03-07T07:12:41.898188+0000 mon.vm09 (mon.1) 52 : audit [DBG] from='client.? 192.168.123.107:0/3872206163' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:43.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:43 vm09 bash[21220]: audit 2026-03-07T07:12:41.707116+0000 mgr.vm07.yrfcuj (mgr.14201) 989 : audit [DBG] from='client.25693 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:43.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:43 vm09 bash[21220]: audit 2026-03-07T07:12:41.707116+0000 mgr.vm07.yrfcuj (mgr.14201) 989 : audit [DBG] from='client.25693 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:43.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:43 vm09 bash[21220]: cluster 2026-03-07T07:12:41.950614+0000 mgr.vm07.yrfcuj (mgr.14201) 990 : cluster [DBG] pgmap v584: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:43.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:43 vm09 bash[21220]: cluster 2026-03-07T07:12:41.950614+0000 mgr.vm07.yrfcuj (mgr.14201) 990 : cluster [DBG] pgmap v584: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:43.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:43 vm07 bash[17031]: audit 2026-03-07T07:12:41.707116+0000 mgr.vm07.yrfcuj (mgr.14201) 989 : audit [DBG] from='client.25693 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:43.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:43 vm07 bash[17031]: audit 2026-03-07T07:12:41.707116+0000 mgr.vm07.yrfcuj (mgr.14201) 989 : audit [DBG] from='client.25693 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:43.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:43 vm07 bash[17031]: cluster 2026-03-07T07:12:41.950614+0000 mgr.vm07.yrfcuj (mgr.14201) 990 : cluster [DBG] pgmap v584: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:43.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:43 vm07 bash[17031]: cluster 2026-03-07T07:12:41.950614+0000 mgr.vm07.yrfcuj (mgr.14201) 990 : cluster [DBG] pgmap v584: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:12:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:45 vm09 bash[21220]: cluster 2026-03-07T07:12:43.951206+0000 mgr.vm07.yrfcuj (mgr.14201) 991 : cluster [DBG] pgmap v585: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:45 vm09 bash[21220]: cluster 2026-03-07T07:12:43.951206+0000 mgr.vm07.yrfcuj (mgr.14201) 991 : cluster [DBG] pgmap v585: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:45.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:45 vm07 bash[17031]: cluster 2026-03-07T07:12:43.951206+0000 mgr.vm07.yrfcuj (mgr.14201) 991 : cluster [DBG] pgmap v585: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:45.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:45 vm07 bash[17031]: cluster 2026-03-07T07:12:43.951206+0000 mgr.vm07.yrfcuj (mgr.14201) 991 : cluster [DBG] pgmap v585: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:47.094 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:12:47.265 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:12:47.265 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (9m) 8m ago 15m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:12:47.265 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (8m) 8m ago 15m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:12:47.265 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 2m ago 15m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:12:47.265 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 2m ago 15m - - 2026-03-07T08:12:47.462 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:12:47.462 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:12:47.462 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:12:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:47 vm09 bash[21220]: cluster 2026-03-07T07:12:45.951577+0000 mgr.vm07.yrfcuj (mgr.14201) 992 : cluster [DBG] pgmap v586: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:47 vm09 bash[21220]: cluster 2026-03-07T07:12:45.951577+0000 mgr.vm07.yrfcuj (mgr.14201) 992 : cluster [DBG] pgmap v586: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:47 vm09 bash[21220]: audit 2026-03-07T07:12:47.466904+0000 mon.vm07 (mon.0) 1193 : audit [DBG] from='client.? 192.168.123.107:0/1361365417' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:47 vm09 bash[21220]: audit 2026-03-07T07:12:47.466904+0000 mon.vm07 (mon.0) 1193 : audit [DBG] from='client.? 192.168.123.107:0/1361365417' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:47.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:47 vm07 bash[17031]: cluster 2026-03-07T07:12:45.951577+0000 mgr.vm07.yrfcuj (mgr.14201) 992 : cluster [DBG] pgmap v586: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:47.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:47 vm07 bash[17031]: cluster 2026-03-07T07:12:45.951577+0000 mgr.vm07.yrfcuj (mgr.14201) 992 : cluster [DBG] pgmap v586: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:47.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:47 vm07 bash[17031]: audit 2026-03-07T07:12:47.466904+0000 mon.vm07 (mon.0) 1193 : audit [DBG] from='client.? 192.168.123.107:0/1361365417' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:47.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:47 vm07 bash[17031]: audit 2026-03-07T07:12:47.466904+0000 mon.vm07 (mon.0) 1193 : audit [DBG] from='client.? 192.168.123.107:0/1361365417' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:48 vm09 bash[21220]: audit 2026-03-07T07:12:47.071510+0000 mgr.vm07.yrfcuj (mgr.14201) 993 : audit [DBG] from='client.16598 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:48 vm09 bash[21220]: audit 2026-03-07T07:12:47.071510+0000 mgr.vm07.yrfcuj (mgr.14201) 993 : audit [DBG] from='client.16598 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:48 vm09 bash[21220]: audit 2026-03-07T07:12:47.267255+0000 mgr.vm07.yrfcuj (mgr.14201) 994 : audit [DBG] from='client.16602 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:48 vm09 bash[21220]: audit 2026-03-07T07:12:47.267255+0000 mgr.vm07.yrfcuj (mgr.14201) 994 : audit [DBG] from='client.16602 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:48.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:48 vm07 bash[17031]: audit 2026-03-07T07:12:47.071510+0000 mgr.vm07.yrfcuj (mgr.14201) 993 : audit [DBG] from='client.16598 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:48.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:48 vm07 bash[17031]: audit 2026-03-07T07:12:47.071510+0000 mgr.vm07.yrfcuj (mgr.14201) 993 : audit [DBG] from='client.16598 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:48.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:48 vm07 bash[17031]: audit 2026-03-07T07:12:47.267255+0000 mgr.vm07.yrfcuj (mgr.14201) 994 : audit [DBG] from='client.16602 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:48.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:48 vm07 bash[17031]: audit 2026-03-07T07:12:47.267255+0000 mgr.vm07.yrfcuj (mgr.14201) 994 : audit [DBG] from='client.16602 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:49 vm09 bash[21220]: cluster 2026-03-07T07:12:47.952230+0000 mgr.vm07.yrfcuj (mgr.14201) 995 : cluster [DBG] pgmap v587: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:49 vm09 bash[21220]: cluster 2026-03-07T07:12:47.952230+0000 mgr.vm07.yrfcuj (mgr.14201) 995 : cluster [DBG] pgmap v587: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:49.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:49 vm07 bash[17031]: cluster 2026-03-07T07:12:47.952230+0000 mgr.vm07.yrfcuj (mgr.14201) 995 : cluster [DBG] pgmap v587: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:49.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:49 vm07 bash[17031]: cluster 2026-03-07T07:12:47.952230+0000 mgr.vm07.yrfcuj (mgr.14201) 995 : cluster [DBG] pgmap v587: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:51 vm09 bash[21220]: cluster 2026-03-07T07:12:49.952741+0000 mgr.vm07.yrfcuj (mgr.14201) 996 : cluster [DBG] pgmap v588: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:51 vm09 bash[21220]: cluster 2026-03-07T07:12:49.952741+0000 mgr.vm07.yrfcuj (mgr.14201) 996 : cluster [DBG] pgmap v588: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:51 vm07 bash[17031]: cluster 2026-03-07T07:12:49.952741+0000 mgr.vm07.yrfcuj (mgr.14201) 996 : cluster [DBG] pgmap v588: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:51.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:51 vm07 bash[17031]: cluster 2026-03-07T07:12:49.952741+0000 mgr.vm07.yrfcuj (mgr.14201) 996 : cluster [DBG] pgmap v588: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:52.649 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:12:52.813 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:12:52.813 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (9m) 8m ago 15m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:12:52.813 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (8m) 8m ago 15m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:12:52.813 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 2m ago 15m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:12:52.813 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 2m ago 15m - - 2026-03-07T08:12:53.015 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:12:53.015 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:12:53.015 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:12:53.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:53 vm09 bash[21220]: cluster 2026-03-07T07:12:51.953225+0000 mgr.vm07.yrfcuj (mgr.14201) 997 : cluster [DBG] pgmap v589: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:53.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:53 vm09 bash[21220]: cluster 2026-03-07T07:12:51.953225+0000 mgr.vm07.yrfcuj (mgr.14201) 997 : cluster [DBG] pgmap v589: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:53.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:53 vm09 bash[21220]: audit 2026-03-07T07:12:53.020070+0000 mon.vm07 (mon.0) 1194 : audit [DBG] from='client.? 192.168.123.107:0/3540110788' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:53.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:53 vm09 bash[21220]: audit 2026-03-07T07:12:53.020070+0000 mon.vm07 (mon.0) 1194 : audit [DBG] from='client.? 192.168.123.107:0/3540110788' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:53.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:53 vm07 bash[17031]: cluster 2026-03-07T07:12:51.953225+0000 mgr.vm07.yrfcuj (mgr.14201) 997 : cluster [DBG] pgmap v589: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:53.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:53 vm07 bash[17031]: cluster 2026-03-07T07:12:51.953225+0000 mgr.vm07.yrfcuj (mgr.14201) 997 : cluster [DBG] pgmap v589: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:12:53.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:53 vm07 bash[17031]: audit 2026-03-07T07:12:53.020070+0000 mon.vm07 (mon.0) 1194 : audit [DBG] from='client.? 192.168.123.107:0/3540110788' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:53.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:53 vm07 bash[17031]: audit 2026-03-07T07:12:53.020070+0000 mon.vm07 (mon.0) 1194 : audit [DBG] from='client.? 192.168.123.107:0/3540110788' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:54 vm09 bash[21220]: audit 2026-03-07T07:12:52.637559+0000 mgr.vm07.yrfcuj (mgr.14201) 998 : audit [DBG] from='client.16610 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:54 vm09 bash[21220]: audit 2026-03-07T07:12:52.637559+0000 mgr.vm07.yrfcuj (mgr.14201) 998 : audit [DBG] from='client.16610 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:54 vm09 bash[21220]: audit 2026-03-07T07:12:52.815244+0000 mgr.vm07.yrfcuj (mgr.14201) 999 : audit [DBG] from='client.16614 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:54 vm09 bash[21220]: audit 2026-03-07T07:12:52.815244+0000 mgr.vm07.yrfcuj (mgr.14201) 999 : audit [DBG] from='client.16614 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:54 vm09 bash[21220]: audit 2026-03-07T07:12:53.992879+0000 mon.vm07 (mon.0) 1195 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:12:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:54 vm09 bash[21220]: audit 2026-03-07T07:12:53.992879+0000 mon.vm07 (mon.0) 1195 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:12:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:54 vm09 bash[21220]: audit 2026-03-07T07:12:54.330576+0000 mon.vm07 (mon.0) 1196 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:12:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:54 vm09 bash[21220]: audit 2026-03-07T07:12:54.330576+0000 mon.vm07 (mon.0) 1196 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:12:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:54 vm09 bash[21220]: audit 2026-03-07T07:12:54.331137+0000 mon.vm07 (mon.0) 1197 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:12:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:54 vm09 bash[21220]: audit 2026-03-07T07:12:54.331137+0000 mon.vm07 (mon.0) 1197 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:12:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:54 vm09 bash[21220]: audit 2026-03-07T07:12:54.337033+0000 mon.vm07 (mon.0) 1198 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:12:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:54 vm09 bash[21220]: audit 2026-03-07T07:12:54.337033+0000 mon.vm07 (mon.0) 1198 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:12:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:54 vm09 bash[21220]: audit 2026-03-07T07:12:54.338909+0000 mon.vm07 (mon.0) 1199 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:12:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:54 vm09 bash[21220]: audit 2026-03-07T07:12:54.338909+0000 mon.vm07 (mon.0) 1199 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:12:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:54 vm07 bash[17031]: audit 2026-03-07T07:12:52.637559+0000 mgr.vm07.yrfcuj (mgr.14201) 998 : audit [DBG] from='client.16610 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:54 vm07 bash[17031]: audit 2026-03-07T07:12:52.637559+0000 mgr.vm07.yrfcuj (mgr.14201) 998 : audit [DBG] from='client.16610 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:54 vm07 bash[17031]: audit 2026-03-07T07:12:52.815244+0000 mgr.vm07.yrfcuj (mgr.14201) 999 : audit [DBG] from='client.16614 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:54 vm07 bash[17031]: audit 2026-03-07T07:12:52.815244+0000 mgr.vm07.yrfcuj (mgr.14201) 999 : audit [DBG] from='client.16614 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:54 vm07 bash[17031]: audit 2026-03-07T07:12:53.992879+0000 mon.vm07 (mon.0) 1195 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:12:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:54 vm07 bash[17031]: audit 2026-03-07T07:12:53.992879+0000 mon.vm07 (mon.0) 1195 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:12:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:54 vm07 bash[17031]: audit 2026-03-07T07:12:54.330576+0000 mon.vm07 (mon.0) 1196 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:12:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:54 vm07 bash[17031]: audit 2026-03-07T07:12:54.330576+0000 mon.vm07 (mon.0) 1196 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:12:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:54 vm07 bash[17031]: audit 2026-03-07T07:12:54.331137+0000 mon.vm07 (mon.0) 1197 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:12:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:54 vm07 bash[17031]: audit 2026-03-07T07:12:54.331137+0000 mon.vm07 (mon.0) 1197 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:12:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:54 vm07 bash[17031]: audit 2026-03-07T07:12:54.337033+0000 mon.vm07 (mon.0) 1198 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:12:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:54 vm07 bash[17031]: audit 2026-03-07T07:12:54.337033+0000 mon.vm07 (mon.0) 1198 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:12:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:54 vm07 bash[17031]: audit 2026-03-07T07:12:54.338909+0000 mon.vm07 (mon.0) 1199 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:12:54.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:54 vm07 bash[17031]: audit 2026-03-07T07:12:54.338909+0000 mon.vm07 (mon.0) 1199 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:12:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:55 vm09 bash[21220]: cluster 2026-03-07T07:12:53.953667+0000 mgr.vm07.yrfcuj (mgr.14201) 1000 : cluster [DBG] pgmap v590: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:55 vm09 bash[21220]: cluster 2026-03-07T07:12:53.953667+0000 mgr.vm07.yrfcuj (mgr.14201) 1000 : cluster [DBG] pgmap v590: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:55 vm09 bash[21220]: cluster 2026-03-07T07:12:54.331959+0000 mgr.vm07.yrfcuj (mgr.14201) 1001 : cluster [DBG] pgmap v591: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:55 vm09 bash[21220]: cluster 2026-03-07T07:12:54.331959+0000 mgr.vm07.yrfcuj (mgr.14201) 1001 : cluster [DBG] pgmap v591: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:55 vm09 bash[21220]: cluster 2026-03-07T07:12:54.332134+0000 mgr.vm07.yrfcuj (mgr.14201) 1002 : cluster [DBG] pgmap v592: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:55 vm09 bash[21220]: cluster 2026-03-07T07:12:54.332134+0000 mgr.vm07.yrfcuj (mgr.14201) 1002 : cluster [DBG] pgmap v592: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:55 vm09 bash[21220]: audit 2026-03-07T07:12:55.144262+0000 mon.vm07 (mon.0) 1200 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:12:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:55 vm09 bash[21220]: audit 2026-03-07T07:12:55.144262+0000 mon.vm07 (mon.0) 1200 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:12:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:55 vm07 bash[17031]: cluster 2026-03-07T07:12:53.953667+0000 mgr.vm07.yrfcuj (mgr.14201) 1000 : cluster [DBG] pgmap v590: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:55 vm07 bash[17031]: cluster 2026-03-07T07:12:53.953667+0000 mgr.vm07.yrfcuj (mgr.14201) 1000 : cluster [DBG] pgmap v590: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:55 vm07 bash[17031]: cluster 2026-03-07T07:12:54.331959+0000 mgr.vm07.yrfcuj (mgr.14201) 1001 : cluster [DBG] pgmap v591: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:55 vm07 bash[17031]: cluster 2026-03-07T07:12:54.331959+0000 mgr.vm07.yrfcuj (mgr.14201) 1001 : cluster [DBG] pgmap v591: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:55 vm07 bash[17031]: cluster 2026-03-07T07:12:54.332134+0000 mgr.vm07.yrfcuj (mgr.14201) 1002 : cluster [DBG] pgmap v592: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:55 vm07 bash[17031]: cluster 2026-03-07T07:12:54.332134+0000 mgr.vm07.yrfcuj (mgr.14201) 1002 : cluster [DBG] pgmap v592: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:55 vm07 bash[17031]: audit 2026-03-07T07:12:55.144262+0000 mon.vm07 (mon.0) 1200 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:12:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:55 vm07 bash[17031]: audit 2026-03-07T07:12:55.144262+0000 mon.vm07 (mon.0) 1200 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:12:57.844 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:57 vm07 bash[17031]: cluster 2026-03-07T07:12:56.332539+0000 mgr.vm07.yrfcuj (mgr.14201) 1003 : cluster [DBG] pgmap v593: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:57.844 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:57 vm07 bash[17031]: cluster 2026-03-07T07:12:56.332539+0000 mgr.vm07.yrfcuj (mgr.14201) 1003 : cluster [DBG] pgmap v593: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:57.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:57 vm09 bash[21220]: cluster 2026-03-07T07:12:56.332539+0000 mgr.vm07.yrfcuj (mgr.14201) 1003 : cluster [DBG] pgmap v593: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:57.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:57 vm09 bash[21220]: cluster 2026-03-07T07:12:56.332539+0000 mgr.vm07.yrfcuj (mgr.14201) 1003 : cluster [DBG] pgmap v593: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:12:58.210 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:12:58.367 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:12:58.367 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (9m) 8m ago 15m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:12:58.367 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (8m) 8m ago 15m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:12:58.368 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 3m ago 15m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:12:58.368 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 3m ago 15m - - 2026-03-07T08:12:58.568 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:12:58.568 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:12:58.568 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:12:58.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:58 vm09 bash[21220]: audit 2026-03-07T07:12:58.573040+0000 mon.vm07 (mon.0) 1201 : audit [DBG] from='client.? 192.168.123.107:0/1792198334' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:58.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:58 vm09 bash[21220]: audit 2026-03-07T07:12:58.573040+0000 mon.vm07 (mon.0) 1201 : audit [DBG] from='client.? 192.168.123.107:0/1792198334' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:58.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:58 vm07 bash[17031]: audit 2026-03-07T07:12:58.573040+0000 mon.vm07 (mon.0) 1201 : audit [DBG] from='client.? 192.168.123.107:0/1792198334' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:58.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:58 vm07 bash[17031]: audit 2026-03-07T07:12:58.573040+0000 mon.vm07 (mon.0) 1201 : audit [DBG] from='client.? 192.168.123.107:0/1792198334' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:12:59.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:59 vm09 bash[21220]: audit 2026-03-07T07:12:58.200569+0000 mgr.vm07.yrfcuj (mgr.14201) 1004 : audit [DBG] from='client.16622 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:59.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:59 vm09 bash[21220]: audit 2026-03-07T07:12:58.200569+0000 mgr.vm07.yrfcuj (mgr.14201) 1004 : audit [DBG] from='client.16622 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:59.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:59 vm09 bash[21220]: cluster 2026-03-07T07:12:58.333041+0000 mgr.vm07.yrfcuj (mgr.14201) 1005 : cluster [DBG] pgmap v594: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:12:59.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:59 vm09 bash[21220]: cluster 2026-03-07T07:12:58.333041+0000 mgr.vm07.yrfcuj (mgr.14201) 1005 : cluster [DBG] pgmap v594: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:12:59.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:59 vm09 bash[21220]: audit 2026-03-07T07:12:58.368852+0000 mgr.vm07.yrfcuj (mgr.14201) 1006 : audit [DBG] from='client.16626 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:59.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:12:59 vm09 bash[21220]: audit 2026-03-07T07:12:58.368852+0000 mgr.vm07.yrfcuj (mgr.14201) 1006 : audit [DBG] from='client.16626 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:59.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:59 vm07 bash[17031]: audit 2026-03-07T07:12:58.200569+0000 mgr.vm07.yrfcuj (mgr.14201) 1004 : audit [DBG] from='client.16622 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:59.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:59 vm07 bash[17031]: audit 2026-03-07T07:12:58.200569+0000 mgr.vm07.yrfcuj (mgr.14201) 1004 : audit [DBG] from='client.16622 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:59.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:59 vm07 bash[17031]: cluster 2026-03-07T07:12:58.333041+0000 mgr.vm07.yrfcuj (mgr.14201) 1005 : cluster [DBG] pgmap v594: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:12:59.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:59 vm07 bash[17031]: cluster 2026-03-07T07:12:58.333041+0000 mgr.vm07.yrfcuj (mgr.14201) 1005 : cluster [DBG] pgmap v594: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:12:59.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:59 vm07 bash[17031]: audit 2026-03-07T07:12:58.368852+0000 mgr.vm07.yrfcuj (mgr.14201) 1006 : audit [DBG] from='client.16626 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:12:59.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:12:59 vm07 bash[17031]: audit 2026-03-07T07:12:58.368852+0000 mgr.vm07.yrfcuj (mgr.14201) 1006 : audit [DBG] from='client.16626 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:01 vm09 bash[21220]: cluster 2026-03-07T07:13:00.333508+0000 mgr.vm07.yrfcuj (mgr.14201) 1007 : cluster [DBG] pgmap v595: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:13:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:01 vm09 bash[21220]: cluster 2026-03-07T07:13:00.333508+0000 mgr.vm07.yrfcuj (mgr.14201) 1007 : cluster [DBG] pgmap v595: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:13:01.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:01 vm07 bash[17031]: cluster 2026-03-07T07:13:00.333508+0000 mgr.vm07.yrfcuj (mgr.14201) 1007 : cluster [DBG] pgmap v595: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:13:01.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:01 vm07 bash[17031]: cluster 2026-03-07T07:13:00.333508+0000 mgr.vm07.yrfcuj (mgr.14201) 1007 : cluster [DBG] pgmap v595: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 122 B/s rd, 244 B/s wr, 0 op/s 2026-03-07T08:13:03.755 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:13:03.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:03 vm09 bash[21220]: cluster 2026-03-07T07:13:02.333929+0000 mgr.vm07.yrfcuj (mgr.14201) 1008 : cluster [DBG] pgmap v596: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 244 B/s rd, 488 B/s wr, 0 op/s 2026-03-07T08:13:03.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:03 vm09 bash[21220]: cluster 2026-03-07T07:13:02.333929+0000 mgr.vm07.yrfcuj (mgr.14201) 1008 : cluster [DBG] pgmap v596: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 244 B/s rd, 488 B/s wr, 0 op/s 2026-03-07T08:13:03.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:03 vm07 bash[17031]: cluster 2026-03-07T07:13:02.333929+0000 mgr.vm07.yrfcuj (mgr.14201) 1008 : cluster [DBG] pgmap v596: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 244 B/s rd, 488 B/s wr, 0 op/s 2026-03-07T08:13:03.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:03 vm07 bash[17031]: cluster 2026-03-07T07:13:02.333929+0000 mgr.vm07.yrfcuj (mgr.14201) 1008 : cluster [DBG] pgmap v596: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 244 B/s rd, 488 B/s wr, 0 op/s 2026-03-07T08:13:03.913 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:13:03.913 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (9m) 8m ago 15m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:13:03.913 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (8m) 8m ago 15m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:13:03.913 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 3m ago 15m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:13:03.913 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 3m ago 15m - - 2026-03-07T08:13:04.108 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:13:04.108 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:13:04.108 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:13:04.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:04 vm07 bash[17031]: audit 2026-03-07T07:13:04.113279+0000 mon.vm07 (mon.0) 1202 : audit [DBG] from='client.? 192.168.123.107:0/790842558' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:04.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:04 vm07 bash[17031]: audit 2026-03-07T07:13:04.113279+0000 mon.vm07 (mon.0) 1202 : audit [DBG] from='client.? 192.168.123.107:0/790842558' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:05.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:04 vm09 bash[21220]: audit 2026-03-07T07:13:04.113279+0000 mon.vm07 (mon.0) 1202 : audit [DBG] from='client.? 192.168.123.107:0/790842558' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:05.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:04 vm09 bash[21220]: audit 2026-03-07T07:13:04.113279+0000 mon.vm07 (mon.0) 1202 : audit [DBG] from='client.? 192.168.123.107:0/790842558' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:05.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:05 vm07 bash[17031]: audit 2026-03-07T07:13:03.743492+0000 mgr.vm07.yrfcuj (mgr.14201) 1009 : audit [DBG] from='client.16634 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:05.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:05 vm07 bash[17031]: audit 2026-03-07T07:13:03.743492+0000 mgr.vm07.yrfcuj (mgr.14201) 1009 : audit [DBG] from='client.16634 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:05.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:05 vm07 bash[17031]: audit 2026-03-07T07:13:03.915158+0000 mgr.vm07.yrfcuj (mgr.14201) 1010 : audit [DBG] from='client.16638 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:05.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:05 vm07 bash[17031]: audit 2026-03-07T07:13:03.915158+0000 mgr.vm07.yrfcuj (mgr.14201) 1010 : audit [DBG] from='client.16638 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:05.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:05 vm07 bash[17031]: cluster 2026-03-07T07:13:04.334417+0000 mgr.vm07.yrfcuj (mgr.14201) 1011 : cluster [DBG] pgmap v597: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 204 B/s rd, 409 B/s wr, 0 op/s 2026-03-07T08:13:05.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:05 vm07 bash[17031]: cluster 2026-03-07T07:13:04.334417+0000 mgr.vm07.yrfcuj (mgr.14201) 1011 : cluster [DBG] pgmap v597: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 204 B/s rd, 409 B/s wr, 0 op/s 2026-03-07T08:13:06.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:05 vm09 bash[21220]: audit 2026-03-07T07:13:03.743492+0000 mgr.vm07.yrfcuj (mgr.14201) 1009 : audit [DBG] from='client.16634 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:06.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:05 vm09 bash[21220]: audit 2026-03-07T07:13:03.743492+0000 mgr.vm07.yrfcuj (mgr.14201) 1009 : audit [DBG] from='client.16634 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:06.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:05 vm09 bash[21220]: audit 2026-03-07T07:13:03.915158+0000 mgr.vm07.yrfcuj (mgr.14201) 1010 : audit [DBG] from='client.16638 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:06.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:05 vm09 bash[21220]: audit 2026-03-07T07:13:03.915158+0000 mgr.vm07.yrfcuj (mgr.14201) 1010 : audit [DBG] from='client.16638 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:06.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:05 vm09 bash[21220]: cluster 2026-03-07T07:13:04.334417+0000 mgr.vm07.yrfcuj (mgr.14201) 1011 : cluster [DBG] pgmap v597: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 204 B/s rd, 409 B/s wr, 0 op/s 2026-03-07T08:13:06.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:05 vm09 bash[21220]: cluster 2026-03-07T07:13:04.334417+0000 mgr.vm07.yrfcuj (mgr.14201) 1011 : cluster [DBG] pgmap v597: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 204 B/s rd, 409 B/s wr, 0 op/s 2026-03-07T08:13:07.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:06 vm09 bash[21220]: cluster 2026-03-07T07:13:06.334761+0000 mgr.vm07.yrfcuj (mgr.14201) 1012 : cluster [DBG] pgmap v598: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:13:07.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:06 vm09 bash[21220]: cluster 2026-03-07T07:13:06.334761+0000 mgr.vm07.yrfcuj (mgr.14201) 1012 : cluster [DBG] pgmap v598: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:13:07.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:06 vm07 bash[17031]: cluster 2026-03-07T07:13:06.334761+0000 mgr.vm07.yrfcuj (mgr.14201) 1012 : cluster [DBG] pgmap v598: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:13:07.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:06 vm07 bash[17031]: cluster 2026-03-07T07:13:06.334761+0000 mgr.vm07.yrfcuj (mgr.14201) 1012 : cluster [DBG] pgmap v598: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:13:09.298 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:13:09.460 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:13:09.461 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (9m) 8m ago 15m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:13:09.461 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (9m) 8m ago 15m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:13:09.461 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 3m ago 15m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:13:09.461 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 3m ago 15m - - 2026-03-07T08:13:09.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:09 vm07 bash[17031]: cluster 2026-03-07T07:13:08.335388+0000 mgr.vm07.yrfcuj (mgr.14201) 1013 : cluster [DBG] pgmap v599: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:13:09.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:09 vm07 bash[17031]: cluster 2026-03-07T07:13:08.335388+0000 mgr.vm07.yrfcuj (mgr.14201) 1013 : cluster [DBG] pgmap v599: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:13:09.663 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:13:09.663 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:13:09.663 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:13:09.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:09 vm09 bash[21220]: cluster 2026-03-07T07:13:08.335388+0000 mgr.vm07.yrfcuj (mgr.14201) 1013 : cluster [DBG] pgmap v599: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:13:09.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:09 vm09 bash[21220]: cluster 2026-03-07T07:13:08.335388+0000 mgr.vm07.yrfcuj (mgr.14201) 1013 : cluster [DBG] pgmap v599: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:13:10.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:10 vm07 bash[17031]: audit 2026-03-07T07:13:09.286349+0000 mgr.vm07.yrfcuj (mgr.14201) 1014 : audit [DBG] from='client.16646 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:10.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:10 vm07 bash[17031]: audit 2026-03-07T07:13:09.286349+0000 mgr.vm07.yrfcuj (mgr.14201) 1014 : audit [DBG] from='client.16646 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:10.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:10 vm07 bash[17031]: audit 2026-03-07T07:13:09.461087+0000 mgr.vm07.yrfcuj (mgr.14201) 1015 : audit [DBG] from='client.16650 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:10.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:10 vm07 bash[17031]: audit 2026-03-07T07:13:09.461087+0000 mgr.vm07.yrfcuj (mgr.14201) 1015 : audit [DBG] from='client.16650 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:10.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:10 vm07 bash[17031]: audit 2026-03-07T07:13:09.668132+0000 mon.vm07 (mon.0) 1203 : audit [DBG] from='client.? 192.168.123.107:0/1464408251' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:10.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:10 vm07 bash[17031]: audit 2026-03-07T07:13:09.668132+0000 mon.vm07 (mon.0) 1203 : audit [DBG] from='client.? 192.168.123.107:0/1464408251' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:10.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:10 vm07 bash[17031]: audit 2026-03-07T07:13:10.144472+0000 mon.vm07 (mon.0) 1204 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:13:10.648 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:10 vm07 bash[17031]: audit 2026-03-07T07:13:10.144472+0000 mon.vm07 (mon.0) 1204 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:13:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:10 vm09 bash[21220]: audit 2026-03-07T07:13:09.286349+0000 mgr.vm07.yrfcuj (mgr.14201) 1014 : audit [DBG] from='client.16646 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:10 vm09 bash[21220]: audit 2026-03-07T07:13:09.286349+0000 mgr.vm07.yrfcuj (mgr.14201) 1014 : audit [DBG] from='client.16646 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:10 vm09 bash[21220]: audit 2026-03-07T07:13:09.461087+0000 mgr.vm07.yrfcuj (mgr.14201) 1015 : audit [DBG] from='client.16650 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:10 vm09 bash[21220]: audit 2026-03-07T07:13:09.461087+0000 mgr.vm07.yrfcuj (mgr.14201) 1015 : audit [DBG] from='client.16650 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:10 vm09 bash[21220]: audit 2026-03-07T07:13:09.668132+0000 mon.vm07 (mon.0) 1203 : audit [DBG] from='client.? 192.168.123.107:0/1464408251' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:10 vm09 bash[21220]: audit 2026-03-07T07:13:09.668132+0000 mon.vm07 (mon.0) 1203 : audit [DBG] from='client.? 192.168.123.107:0/1464408251' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:10 vm09 bash[21220]: audit 2026-03-07T07:13:10.144472+0000 mon.vm07 (mon.0) 1204 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:13:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:10 vm09 bash[21220]: audit 2026-03-07T07:13:10.144472+0000 mon.vm07 (mon.0) 1204 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:13:11.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:11 vm09 bash[21220]: cluster 2026-03-07T07:13:10.335839+0000 mgr.vm07.yrfcuj (mgr.14201) 1016 : cluster [DBG] pgmap v600: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:11.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:11 vm09 bash[21220]: cluster 2026-03-07T07:13:10.335839+0000 mgr.vm07.yrfcuj (mgr.14201) 1016 : cluster [DBG] pgmap v600: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:11.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:11 vm07 bash[17031]: cluster 2026-03-07T07:13:10.335839+0000 mgr.vm07.yrfcuj (mgr.14201) 1016 : cluster [DBG] pgmap v600: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:11.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:11 vm07 bash[17031]: cluster 2026-03-07T07:13:10.335839+0000 mgr.vm07.yrfcuj (mgr.14201) 1016 : cluster [DBG] pgmap v600: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:13.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:13 vm09 bash[21220]: cluster 2026-03-07T07:13:12.336263+0000 mgr.vm07.yrfcuj (mgr.14201) 1017 : cluster [DBG] pgmap v601: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:13:13.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:13 vm09 bash[21220]: cluster 2026-03-07T07:13:12.336263+0000 mgr.vm07.yrfcuj (mgr.14201) 1017 : cluster [DBG] pgmap v601: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:13:13.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:13 vm07 bash[17031]: cluster 2026-03-07T07:13:12.336263+0000 mgr.vm07.yrfcuj (mgr.14201) 1017 : cluster [DBG] pgmap v601: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:13:13.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:13 vm07 bash[17031]: cluster 2026-03-07T07:13:12.336263+0000 mgr.vm07.yrfcuj (mgr.14201) 1017 : cluster [DBG] pgmap v601: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:13:14.850 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:13:15.015 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:13:15.016 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (9m) 9m ago 15m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:13:15.016 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (9m) 9m ago 15m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:13:15.016 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 3m ago 15m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:13:15.016 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 3m ago 15m - - 2026-03-07T08:13:15.233 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:13:15.233 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:13:15.233 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:13:15.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:15 vm09 bash[21220]: cluster 2026-03-07T07:13:14.336681+0000 mgr.vm07.yrfcuj (mgr.14201) 1018 : cluster [DBG] pgmap v602: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:15.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:15 vm09 bash[21220]: cluster 2026-03-07T07:13:14.336681+0000 mgr.vm07.yrfcuj (mgr.14201) 1018 : cluster [DBG] pgmap v602: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:15.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:15 vm09 bash[21220]: audit 2026-03-07T07:13:15.238168+0000 mon.vm07 (mon.0) 1205 : audit [DBG] from='client.? 192.168.123.107:0/3913485760' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:15.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:15 vm09 bash[21220]: audit 2026-03-07T07:13:15.238168+0000 mon.vm07 (mon.0) 1205 : audit [DBG] from='client.? 192.168.123.107:0/3913485760' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:15.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:15 vm07 bash[17031]: cluster 2026-03-07T07:13:14.336681+0000 mgr.vm07.yrfcuj (mgr.14201) 1018 : cluster [DBG] pgmap v602: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:15.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:15 vm07 bash[17031]: cluster 2026-03-07T07:13:14.336681+0000 mgr.vm07.yrfcuj (mgr.14201) 1018 : cluster [DBG] pgmap v602: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:15.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:15 vm07 bash[17031]: audit 2026-03-07T07:13:15.238168+0000 mon.vm07 (mon.0) 1205 : audit [DBG] from='client.? 192.168.123.107:0/3913485760' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:15.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:15 vm07 bash[17031]: audit 2026-03-07T07:13:15.238168+0000 mon.vm07 (mon.0) 1205 : audit [DBG] from='client.? 192.168.123.107:0/3913485760' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:16.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:16 vm09 bash[21220]: audit 2026-03-07T07:13:14.839048+0000 mgr.vm07.yrfcuj (mgr.14201) 1019 : audit [DBG] from='client.16658 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:16.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:16 vm09 bash[21220]: audit 2026-03-07T07:13:14.839048+0000 mgr.vm07.yrfcuj (mgr.14201) 1019 : audit [DBG] from='client.16658 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:16.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:16 vm09 bash[21220]: audit 2026-03-07T07:13:15.017191+0000 mgr.vm07.yrfcuj (mgr.14201) 1020 : audit [DBG] from='client.16662 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:16.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:16 vm09 bash[21220]: audit 2026-03-07T07:13:15.017191+0000 mgr.vm07.yrfcuj (mgr.14201) 1020 : audit [DBG] from='client.16662 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:16.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:16 vm07 bash[17031]: audit 2026-03-07T07:13:14.839048+0000 mgr.vm07.yrfcuj (mgr.14201) 1019 : audit [DBG] from='client.16658 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:16.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:16 vm07 bash[17031]: audit 2026-03-07T07:13:14.839048+0000 mgr.vm07.yrfcuj (mgr.14201) 1019 : audit [DBG] from='client.16658 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:16.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:16 vm07 bash[17031]: audit 2026-03-07T07:13:15.017191+0000 mgr.vm07.yrfcuj (mgr.14201) 1020 : audit [DBG] from='client.16662 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:16.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:16 vm07 bash[17031]: audit 2026-03-07T07:13:15.017191+0000 mgr.vm07.yrfcuj (mgr.14201) 1020 : audit [DBG] from='client.16662 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:17.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:17 vm09 bash[21220]: cluster 2026-03-07T07:13:16.337083+0000 mgr.vm07.yrfcuj (mgr.14201) 1021 : cluster [DBG] pgmap v603: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:17.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:17 vm09 bash[21220]: cluster 2026-03-07T07:13:16.337083+0000 mgr.vm07.yrfcuj (mgr.14201) 1021 : cluster [DBG] pgmap v603: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:17.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:17 vm07 bash[17031]: cluster 2026-03-07T07:13:16.337083+0000 mgr.vm07.yrfcuj (mgr.14201) 1021 : cluster [DBG] pgmap v603: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:17.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:17 vm07 bash[17031]: cluster 2026-03-07T07:13:16.337083+0000 mgr.vm07.yrfcuj (mgr.14201) 1021 : cluster [DBG] pgmap v603: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:19 vm09 bash[21220]: cluster 2026-03-07T07:13:18.337589+0000 mgr.vm07.yrfcuj (mgr.14201) 1022 : cluster [DBG] pgmap v604: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:19.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:19 vm09 bash[21220]: cluster 2026-03-07T07:13:18.337589+0000 mgr.vm07.yrfcuj (mgr.14201) 1022 : cluster [DBG] pgmap v604: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:19.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:19 vm07 bash[17031]: cluster 2026-03-07T07:13:18.337589+0000 mgr.vm07.yrfcuj (mgr.14201) 1022 : cluster [DBG] pgmap v604: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:19.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:19 vm07 bash[17031]: cluster 2026-03-07T07:13:18.337589+0000 mgr.vm07.yrfcuj (mgr.14201) 1022 : cluster [DBG] pgmap v604: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:20.422 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:13:20.587 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:13:20.587 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (9m) 9m ago 15m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:13:20.587 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (9m) 9m ago 15m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:13:20.587 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (3m) 3m ago 15m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:13:20.587 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 3m ago 15m - - 2026-03-07T08:13:20.810 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:13:20.810 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:13:20.810 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:13:21.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:21 vm09 bash[21220]: cluster 2026-03-07T07:13:20.338043+0000 mgr.vm07.yrfcuj (mgr.14201) 1023 : cluster [DBG] pgmap v605: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:21.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:21 vm09 bash[21220]: cluster 2026-03-07T07:13:20.338043+0000 mgr.vm07.yrfcuj (mgr.14201) 1023 : cluster [DBG] pgmap v605: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:21.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:21 vm09 bash[21220]: audit 2026-03-07T07:13:20.411913+0000 mgr.vm07.yrfcuj (mgr.14201) 1024 : audit [DBG] from='client.16670 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:21.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:21 vm09 bash[21220]: audit 2026-03-07T07:13:20.411913+0000 mgr.vm07.yrfcuj (mgr.14201) 1024 : audit [DBG] from='client.16670 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:21.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:21 vm09 bash[21220]: audit 2026-03-07T07:13:20.588811+0000 mgr.vm07.yrfcuj (mgr.14201) 1025 : audit [DBG] from='client.16674 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:21.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:21 vm09 bash[21220]: audit 2026-03-07T07:13:20.588811+0000 mgr.vm07.yrfcuj (mgr.14201) 1025 : audit [DBG] from='client.16674 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:21.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:21 vm09 bash[21220]: audit 2026-03-07T07:13:20.814600+0000 mon.vm07 (mon.0) 1206 : audit [DBG] from='client.? 192.168.123.107:0/402723767' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:21.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:21 vm09 bash[21220]: audit 2026-03-07T07:13:20.814600+0000 mon.vm07 (mon.0) 1206 : audit [DBG] from='client.? 192.168.123.107:0/402723767' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:21.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:21 vm07 bash[17031]: cluster 2026-03-07T07:13:20.338043+0000 mgr.vm07.yrfcuj (mgr.14201) 1023 : cluster [DBG] pgmap v605: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:21.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:21 vm07 bash[17031]: cluster 2026-03-07T07:13:20.338043+0000 mgr.vm07.yrfcuj (mgr.14201) 1023 : cluster [DBG] pgmap v605: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:21.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:21 vm07 bash[17031]: audit 2026-03-07T07:13:20.411913+0000 mgr.vm07.yrfcuj (mgr.14201) 1024 : audit [DBG] from='client.16670 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:21.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:21 vm07 bash[17031]: audit 2026-03-07T07:13:20.411913+0000 mgr.vm07.yrfcuj (mgr.14201) 1024 : audit [DBG] from='client.16670 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:21.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:21 vm07 bash[17031]: audit 2026-03-07T07:13:20.588811+0000 mgr.vm07.yrfcuj (mgr.14201) 1025 : audit [DBG] from='client.16674 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:21.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:21 vm07 bash[17031]: audit 2026-03-07T07:13:20.588811+0000 mgr.vm07.yrfcuj (mgr.14201) 1025 : audit [DBG] from='client.16674 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:21.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:21 vm07 bash[17031]: audit 2026-03-07T07:13:20.814600+0000 mon.vm07 (mon.0) 1206 : audit [DBG] from='client.? 192.168.123.107:0/402723767' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:21.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:21 vm07 bash[17031]: audit 2026-03-07T07:13:20.814600+0000 mon.vm07 (mon.0) 1206 : audit [DBG] from='client.? 192.168.123.107:0/402723767' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:23.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:23 vm09 bash[21220]: cluster 2026-03-07T07:13:22.338458+0000 mgr.vm07.yrfcuj (mgr.14201) 1026 : cluster [DBG] pgmap v606: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:23.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:23 vm09 bash[21220]: cluster 2026-03-07T07:13:22.338458+0000 mgr.vm07.yrfcuj (mgr.14201) 1026 : cluster [DBG] pgmap v606: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:23.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:23 vm07 bash[17031]: cluster 2026-03-07T07:13:22.338458+0000 mgr.vm07.yrfcuj (mgr.14201) 1026 : cluster [DBG] pgmap v606: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:23.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:23 vm07 bash[17031]: cluster 2026-03-07T07:13:22.338458+0000 mgr.vm07.yrfcuj (mgr.14201) 1026 : cluster [DBG] pgmap v606: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:25 vm09 bash[21220]: cluster 2026-03-07T07:13:24.338913+0000 mgr.vm07.yrfcuj (mgr.14201) 1027 : cluster [DBG] pgmap v607: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:25 vm09 bash[21220]: cluster 2026-03-07T07:13:24.338913+0000 mgr.vm07.yrfcuj (mgr.14201) 1027 : cluster [DBG] pgmap v607: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:25 vm09 bash[21220]: audit 2026-03-07T07:13:25.144660+0000 mon.vm07 (mon.0) 1207 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:13:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:25 vm09 bash[21220]: audit 2026-03-07T07:13:25.144660+0000 mon.vm07 (mon.0) 1207 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:13:25.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:25 vm07 bash[17031]: cluster 2026-03-07T07:13:24.338913+0000 mgr.vm07.yrfcuj (mgr.14201) 1027 : cluster [DBG] pgmap v607: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:25.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:25 vm07 bash[17031]: cluster 2026-03-07T07:13:24.338913+0000 mgr.vm07.yrfcuj (mgr.14201) 1027 : cluster [DBG] pgmap v607: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:25.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:25 vm07 bash[17031]: audit 2026-03-07T07:13:25.144660+0000 mon.vm07 (mon.0) 1207 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:13:25.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:25 vm07 bash[17031]: audit 2026-03-07T07:13:25.144660+0000 mon.vm07 (mon.0) 1207 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:13:25.991 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:13:26.147 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:13:26.147 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (9m) 9m ago 15m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:13:26.147 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (9m) 9m ago 15m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:13:26.147 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 3m ago 15m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:13:26.147 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 3m ago 15m - - 2026-03-07T08:13:26.340 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:13:26.340 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:13:26.340 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:13:26.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:26 vm09 bash[21220]: audit 2026-03-07T07:13:26.345111+0000 mon.vm07 (mon.0) 1208 : audit [DBG] from='client.? 192.168.123.107:0/2503806631' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:26.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:26 vm09 bash[21220]: audit 2026-03-07T07:13:26.345111+0000 mon.vm07 (mon.0) 1208 : audit [DBG] from='client.? 192.168.123.107:0/2503806631' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:26.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:26 vm07 bash[17031]: audit 2026-03-07T07:13:26.345111+0000 mon.vm07 (mon.0) 1208 : audit [DBG] from='client.? 192.168.123.107:0/2503806631' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:26.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:26 vm07 bash[17031]: audit 2026-03-07T07:13:26.345111+0000 mon.vm07 (mon.0) 1208 : audit [DBG] from='client.? 192.168.123.107:0/2503806631' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:27.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:27 vm09 bash[21220]: audit 2026-03-07T07:13:25.980300+0000 mgr.vm07.yrfcuj (mgr.14201) 1028 : audit [DBG] from='client.16682 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:27.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:27 vm09 bash[21220]: audit 2026-03-07T07:13:25.980300+0000 mgr.vm07.yrfcuj (mgr.14201) 1028 : audit [DBG] from='client.16682 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:27.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:27 vm09 bash[21220]: audit 2026-03-07T07:13:26.149930+0000 mgr.vm07.yrfcuj (mgr.14201) 1029 : audit [DBG] from='client.16686 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:27.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:27 vm09 bash[21220]: audit 2026-03-07T07:13:26.149930+0000 mgr.vm07.yrfcuj (mgr.14201) 1029 : audit [DBG] from='client.16686 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:27.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:27 vm09 bash[21220]: cluster 2026-03-07T07:13:26.339258+0000 mgr.vm07.yrfcuj (mgr.14201) 1030 : cluster [DBG] pgmap v608: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:27.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:27 vm09 bash[21220]: cluster 2026-03-07T07:13:26.339258+0000 mgr.vm07.yrfcuj (mgr.14201) 1030 : cluster [DBG] pgmap v608: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:27.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:27 vm07 bash[17031]: audit 2026-03-07T07:13:25.980300+0000 mgr.vm07.yrfcuj (mgr.14201) 1028 : audit [DBG] from='client.16682 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:27.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:27 vm07 bash[17031]: audit 2026-03-07T07:13:25.980300+0000 mgr.vm07.yrfcuj (mgr.14201) 1028 : audit [DBG] from='client.16682 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:27.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:27 vm07 bash[17031]: audit 2026-03-07T07:13:26.149930+0000 mgr.vm07.yrfcuj (mgr.14201) 1029 : audit [DBG] from='client.16686 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:27.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:27 vm07 bash[17031]: audit 2026-03-07T07:13:26.149930+0000 mgr.vm07.yrfcuj (mgr.14201) 1029 : audit [DBG] from='client.16686 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:27.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:27 vm07 bash[17031]: cluster 2026-03-07T07:13:26.339258+0000 mgr.vm07.yrfcuj (mgr.14201) 1030 : cluster [DBG] pgmap v608: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:27.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:27 vm07 bash[17031]: cluster 2026-03-07T07:13:26.339258+0000 mgr.vm07.yrfcuj (mgr.14201) 1030 : cluster [DBG] pgmap v608: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:29.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:29 vm09 bash[21220]: cluster 2026-03-07T07:13:28.339694+0000 mgr.vm07.yrfcuj (mgr.14201) 1031 : cluster [DBG] pgmap v609: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:29.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:29 vm09 bash[21220]: cluster 2026-03-07T07:13:28.339694+0000 mgr.vm07.yrfcuj (mgr.14201) 1031 : cluster [DBG] pgmap v609: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:29.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:29 vm07 bash[17031]: cluster 2026-03-07T07:13:28.339694+0000 mgr.vm07.yrfcuj (mgr.14201) 1031 : cluster [DBG] pgmap v609: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:29.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:29 vm07 bash[17031]: cluster 2026-03-07T07:13:28.339694+0000 mgr.vm07.yrfcuj (mgr.14201) 1031 : cluster [DBG] pgmap v609: 129 pgs: 129 active+clean; 454 KiB data, 246 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:31.533 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:13:31.707 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:13:31.708 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (10m) 9m ago 15m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:13:31.708 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (9m) 9m ago 15m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:13:31.708 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 3m ago 15m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:13:31.708 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 3m ago 15m - - 2026-03-07T08:13:31.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:31 vm09 bash[21220]: cluster 2026-03-07T07:13:30.340095+0000 mgr.vm07.yrfcuj (mgr.14201) 1032 : cluster [DBG] pgmap v610: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 6.6 KiB/s rd, 170 B/s wr, 11 op/s 2026-03-07T08:13:31.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:31 vm09 bash[21220]: cluster 2026-03-07T07:13:30.340095+0000 mgr.vm07.yrfcuj (mgr.14201) 1032 : cluster [DBG] pgmap v610: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 6.6 KiB/s rd, 170 B/s wr, 11 op/s 2026-03-07T08:13:31.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:31 vm07 bash[17031]: cluster 2026-03-07T07:13:30.340095+0000 mgr.vm07.yrfcuj (mgr.14201) 1032 : cluster [DBG] pgmap v610: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 6.6 KiB/s rd, 170 B/s wr, 11 op/s 2026-03-07T08:13:31.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:31 vm07 bash[17031]: cluster 2026-03-07T07:13:30.340095+0000 mgr.vm07.yrfcuj (mgr.14201) 1032 : cluster [DBG] pgmap v610: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 6.6 KiB/s rd, 170 B/s wr, 11 op/s 2026-03-07T08:13:31.901 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:13:31.901 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:13:31.901 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:13:32.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:32 vm09 bash[21220]: audit 2026-03-07T07:13:31.524230+0000 mgr.vm07.yrfcuj (mgr.14201) 1033 : audit [DBG] from='client.16694 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:32.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:32 vm09 bash[21220]: audit 2026-03-07T07:13:31.524230+0000 mgr.vm07.yrfcuj (mgr.14201) 1033 : audit [DBG] from='client.16694 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:32.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:32 vm09 bash[21220]: audit 2026-03-07T07:13:31.906123+0000 mon.vm07 (mon.0) 1209 : audit [DBG] from='client.? 192.168.123.107:0/1137271076' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:32.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:32 vm09 bash[21220]: audit 2026-03-07T07:13:31.906123+0000 mon.vm07 (mon.0) 1209 : audit [DBG] from='client.? 192.168.123.107:0/1137271076' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:32.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:32 vm07 bash[17031]: audit 2026-03-07T07:13:31.524230+0000 mgr.vm07.yrfcuj (mgr.14201) 1033 : audit [DBG] from='client.16694 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:32.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:32 vm07 bash[17031]: audit 2026-03-07T07:13:31.524230+0000 mgr.vm07.yrfcuj (mgr.14201) 1033 : audit [DBG] from='client.16694 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:32.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:32 vm07 bash[17031]: audit 2026-03-07T07:13:31.906123+0000 mon.vm07 (mon.0) 1209 : audit [DBG] from='client.? 192.168.123.107:0/1137271076' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:32.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:32 vm07 bash[17031]: audit 2026-03-07T07:13:31.906123+0000 mon.vm07 (mon.0) 1209 : audit [DBG] from='client.? 192.168.123.107:0/1137271076' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:33.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:33 vm09 bash[21220]: audit 2026-03-07T07:13:31.710083+0000 mgr.vm07.yrfcuj (mgr.14201) 1034 : audit [DBG] from='client.16698 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:33.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:33 vm09 bash[21220]: audit 2026-03-07T07:13:31.710083+0000 mgr.vm07.yrfcuj (mgr.14201) 1034 : audit [DBG] from='client.16698 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:33.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:33 vm09 bash[21220]: cluster 2026-03-07T07:13:32.340514+0000 mgr.vm07.yrfcuj (mgr.14201) 1035 : cluster [DBG] pgmap v611: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 20 KiB/s rd, 341 B/s wr, 32 op/s 2026-03-07T08:13:33.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:33 vm09 bash[21220]: cluster 2026-03-07T07:13:32.340514+0000 mgr.vm07.yrfcuj (mgr.14201) 1035 : cluster [DBG] pgmap v611: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 20 KiB/s rd, 341 B/s wr, 32 op/s 2026-03-07T08:13:33.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:33 vm07 bash[17031]: audit 2026-03-07T07:13:31.710083+0000 mgr.vm07.yrfcuj (mgr.14201) 1034 : audit [DBG] from='client.16698 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:33.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:33 vm07 bash[17031]: audit 2026-03-07T07:13:31.710083+0000 mgr.vm07.yrfcuj (mgr.14201) 1034 : audit [DBG] from='client.16698 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:33.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:33 vm07 bash[17031]: cluster 2026-03-07T07:13:32.340514+0000 mgr.vm07.yrfcuj (mgr.14201) 1035 : cluster [DBG] pgmap v611: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 20 KiB/s rd, 341 B/s wr, 32 op/s 2026-03-07T08:13:33.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:33 vm07 bash[17031]: cluster 2026-03-07T07:13:32.340514+0000 mgr.vm07.yrfcuj (mgr.14201) 1035 : cluster [DBG] pgmap v611: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 20 KiB/s rd, 341 B/s wr, 32 op/s 2026-03-07T08:13:35.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:35 vm07 bash[17031]: cluster 2026-03-07T07:13:34.340959+0000 mgr.vm07.yrfcuj (mgr.14201) 1036 : cluster [DBG] pgmap v612: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:13:35.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:35 vm07 bash[17031]: cluster 2026-03-07T07:13:34.340959+0000 mgr.vm07.yrfcuj (mgr.14201) 1036 : cluster [DBG] pgmap v612: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:13:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:35 vm09 bash[21220]: cluster 2026-03-07T07:13:34.340959+0000 mgr.vm07.yrfcuj (mgr.14201) 1036 : cluster [DBG] pgmap v612: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:13:36.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:35 vm09 bash[21220]: cluster 2026-03-07T07:13:34.340959+0000 mgr.vm07.yrfcuj (mgr.14201) 1036 : cluster [DBG] pgmap v612: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:13:37.083 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:13:37.241 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:13:37.242 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (10m) 9m ago 15m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:13:37.242 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (9m) 9m ago 15m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:13:37.242 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 3m ago 16m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:13:37.242 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 3m ago 15m - - 2026-03-07T08:13:37.464 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:13:37.464 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:13:37.464 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:13:37.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:37 vm07 bash[17031]: cluster 2026-03-07T07:13:36.341305+0000 mgr.vm07.yrfcuj (mgr.14201) 1037 : cluster [DBG] pgmap v613: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:13:37.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:37 vm07 bash[17031]: cluster 2026-03-07T07:13:36.341305+0000 mgr.vm07.yrfcuj (mgr.14201) 1037 : cluster [DBG] pgmap v613: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:13:37.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:37 vm07 bash[17031]: audit 2026-03-07T07:13:37.469319+0000 mon.vm07 (mon.0) 1210 : audit [DBG] from='client.? 192.168.123.107:0/657476338' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:37.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:37 vm07 bash[17031]: audit 2026-03-07T07:13:37.469319+0000 mon.vm07 (mon.0) 1210 : audit [DBG] from='client.? 192.168.123.107:0/657476338' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:38.104 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:37 vm09 bash[21220]: cluster 2026-03-07T07:13:36.341305+0000 mgr.vm07.yrfcuj (mgr.14201) 1037 : cluster [DBG] pgmap v613: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:13:38.104 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:37 vm09 bash[21220]: cluster 2026-03-07T07:13:36.341305+0000 mgr.vm07.yrfcuj (mgr.14201) 1037 : cluster [DBG] pgmap v613: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:13:38.104 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:37 vm09 bash[21220]: audit 2026-03-07T07:13:37.469319+0000 mon.vm07 (mon.0) 1210 : audit [DBG] from='client.? 192.168.123.107:0/657476338' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:38.104 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:37 vm09 bash[21220]: audit 2026-03-07T07:13:37.469319+0000 mon.vm07 (mon.0) 1210 : audit [DBG] from='client.? 192.168.123.107:0/657476338' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:38.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:38 vm07 bash[17031]: audit 2026-03-07T07:13:37.071275+0000 mgr.vm07.yrfcuj (mgr.14201) 1038 : audit [DBG] from='client.16706 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:38.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:38 vm07 bash[17031]: audit 2026-03-07T07:13:37.071275+0000 mgr.vm07.yrfcuj (mgr.14201) 1038 : audit [DBG] from='client.16706 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:38.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:38 vm07 bash[17031]: audit 2026-03-07T07:13:37.244210+0000 mgr.vm07.yrfcuj (mgr.14201) 1039 : audit [DBG] from='client.16710 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:38.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:38 vm07 bash[17031]: audit 2026-03-07T07:13:37.244210+0000 mgr.vm07.yrfcuj (mgr.14201) 1039 : audit [DBG] from='client.16710 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:39.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:38 vm09 bash[21220]: audit 2026-03-07T07:13:37.071275+0000 mgr.vm07.yrfcuj (mgr.14201) 1038 : audit [DBG] from='client.16706 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:39.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:38 vm09 bash[21220]: audit 2026-03-07T07:13:37.071275+0000 mgr.vm07.yrfcuj (mgr.14201) 1038 : audit [DBG] from='client.16706 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:39.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:38 vm09 bash[21220]: audit 2026-03-07T07:13:37.244210+0000 mgr.vm07.yrfcuj (mgr.14201) 1039 : audit [DBG] from='client.16710 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:39.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:38 vm09 bash[21220]: audit 2026-03-07T07:13:37.244210+0000 mgr.vm07.yrfcuj (mgr.14201) 1039 : audit [DBG] from='client.16710 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:39.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:39 vm07 bash[17031]: cluster 2026-03-07T07:13:38.341864+0000 mgr.vm07.yrfcuj (mgr.14201) 1040 : cluster [DBG] pgmap v614: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:13:39.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:39 vm07 bash[17031]: cluster 2026-03-07T07:13:38.341864+0000 mgr.vm07.yrfcuj (mgr.14201) 1040 : cluster [DBG] pgmap v614: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:13:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:39 vm09 bash[21220]: cluster 2026-03-07T07:13:38.341864+0000 mgr.vm07.yrfcuj (mgr.14201) 1040 : cluster [DBG] pgmap v614: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:13:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:39 vm09 bash[21220]: cluster 2026-03-07T07:13:38.341864+0000 mgr.vm07.yrfcuj (mgr.14201) 1040 : cluster [DBG] pgmap v614: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:13:40.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:40 vm07 bash[17031]: audit 2026-03-07T07:13:40.144788+0000 mon.vm07 (mon.0) 1211 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:13:40.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:40 vm07 bash[17031]: audit 2026-03-07T07:13:40.144788+0000 mon.vm07 (mon.0) 1211 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:13:40.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:40 vm07 bash[17031]: cluster 2026-03-07T07:13:40.342263+0000 mgr.vm07.yrfcuj (mgr.14201) 1041 : cluster [DBG] pgmap v615: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:13:40.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:40 vm07 bash[17031]: cluster 2026-03-07T07:13:40.342263+0000 mgr.vm07.yrfcuj (mgr.14201) 1041 : cluster [DBG] pgmap v615: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:13:41.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:40 vm09 bash[21220]: audit 2026-03-07T07:13:40.144788+0000 mon.vm07 (mon.0) 1211 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:13:41.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:40 vm09 bash[21220]: audit 2026-03-07T07:13:40.144788+0000 mon.vm07 (mon.0) 1211 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:13:41.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:40 vm09 bash[21220]: cluster 2026-03-07T07:13:40.342263+0000 mgr.vm07.yrfcuj (mgr.14201) 1041 : cluster [DBG] pgmap v615: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:13:41.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:40 vm09 bash[21220]: cluster 2026-03-07T07:13:40.342263+0000 mgr.vm07.yrfcuj (mgr.14201) 1041 : cluster [DBG] pgmap v615: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:13:42.643 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:13:42.799 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:13:42.799 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (10m) 9m ago 16m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:13:42.799 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (9m) 9m ago 16m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:13:42.799 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 3m ago 16m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:13:42.799 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 3m ago 16m - - 2026-03-07T08:13:42.994 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:13:42.994 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:13:42.994 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:13:43.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:43 vm07 bash[17031]: cluster 2026-03-07T07:13:42.342715+0000 mgr.vm07.yrfcuj (mgr.14201) 1042 : cluster [DBG] pgmap v616: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 29 KiB/s rd, 341 B/s wr, 49 op/s 2026-03-07T08:13:43.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:43 vm07 bash[17031]: cluster 2026-03-07T07:13:42.342715+0000 mgr.vm07.yrfcuj (mgr.14201) 1042 : cluster [DBG] pgmap v616: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 29 KiB/s rd, 341 B/s wr, 49 op/s 2026-03-07T08:13:43.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:43 vm07 bash[17031]: audit 2026-03-07T07:13:42.999066+0000 mon.vm07 (mon.0) 1212 : audit [DBG] from='client.? 192.168.123.107:0/3150182164' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:43.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:43 vm07 bash[17031]: audit 2026-03-07T07:13:42.999066+0000 mon.vm07 (mon.0) 1212 : audit [DBG] from='client.? 192.168.123.107:0/3150182164' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:43.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:43 vm09 bash[21220]: cluster 2026-03-07T07:13:42.342715+0000 mgr.vm07.yrfcuj (mgr.14201) 1042 : cluster [DBG] pgmap v616: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 29 KiB/s rd, 341 B/s wr, 49 op/s 2026-03-07T08:13:43.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:43 vm09 bash[21220]: cluster 2026-03-07T07:13:42.342715+0000 mgr.vm07.yrfcuj (mgr.14201) 1042 : cluster [DBG] pgmap v616: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 29 KiB/s rd, 341 B/s wr, 49 op/s 2026-03-07T08:13:43.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:43 vm09 bash[21220]: audit 2026-03-07T07:13:42.999066+0000 mon.vm07 (mon.0) 1212 : audit [DBG] from='client.? 192.168.123.107:0/3150182164' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:43.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:43 vm09 bash[21220]: audit 2026-03-07T07:13:42.999066+0000 mon.vm07 (mon.0) 1212 : audit [DBG] from='client.? 192.168.123.107:0/3150182164' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:44.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:44 vm09 bash[21220]: audit 2026-03-07T07:13:42.632447+0000 mgr.vm07.yrfcuj (mgr.14201) 1043 : audit [DBG] from='client.16718 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:44.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:44 vm09 bash[21220]: audit 2026-03-07T07:13:42.632447+0000 mgr.vm07.yrfcuj (mgr.14201) 1043 : audit [DBG] from='client.16718 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:44.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:44 vm09 bash[21220]: audit 2026-03-07T07:13:42.801346+0000 mgr.vm07.yrfcuj (mgr.14201) 1044 : audit [DBG] from='client.16722 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:44.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:44 vm09 bash[21220]: audit 2026-03-07T07:13:42.801346+0000 mgr.vm07.yrfcuj (mgr.14201) 1044 : audit [DBG] from='client.16722 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:44.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:44 vm07 bash[17031]: audit 2026-03-07T07:13:42.632447+0000 mgr.vm07.yrfcuj (mgr.14201) 1043 : audit [DBG] from='client.16718 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:44.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:44 vm07 bash[17031]: audit 2026-03-07T07:13:42.632447+0000 mgr.vm07.yrfcuj (mgr.14201) 1043 : audit [DBG] from='client.16718 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:44.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:44 vm07 bash[17031]: audit 2026-03-07T07:13:42.801346+0000 mgr.vm07.yrfcuj (mgr.14201) 1044 : audit [DBG] from='client.16722 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:44.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:44 vm07 bash[17031]: audit 2026-03-07T07:13:42.801346+0000 mgr.vm07.yrfcuj (mgr.14201) 1044 : audit [DBG] from='client.16722 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:45 vm09 bash[21220]: cluster 2026-03-07T07:13:44.343150+0000 mgr.vm07.yrfcuj (mgr.14201) 1045 : cluster [DBG] pgmap v617: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 16 KiB/s rd, 170 B/s wr, 27 op/s 2026-03-07T08:13:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:45 vm09 bash[21220]: cluster 2026-03-07T07:13:44.343150+0000 mgr.vm07.yrfcuj (mgr.14201) 1045 : cluster [DBG] pgmap v617: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 16 KiB/s rd, 170 B/s wr, 27 op/s 2026-03-07T08:13:45.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:45 vm07 bash[17031]: cluster 2026-03-07T07:13:44.343150+0000 mgr.vm07.yrfcuj (mgr.14201) 1045 : cluster [DBG] pgmap v617: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 16 KiB/s rd, 170 B/s wr, 27 op/s 2026-03-07T08:13:45.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:45 vm07 bash[17031]: cluster 2026-03-07T07:13:44.343150+0000 mgr.vm07.yrfcuj (mgr.14201) 1045 : cluster [DBG] pgmap v617: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 16 KiB/s rd, 170 B/s wr, 27 op/s 2026-03-07T08:13:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:47 vm09 bash[21220]: cluster 2026-03-07T07:13:46.343527+0000 mgr.vm07.yrfcuj (mgr.14201) 1046 : cluster [DBG] pgmap v618: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:47 vm09 bash[21220]: cluster 2026-03-07T07:13:46.343527+0000 mgr.vm07.yrfcuj (mgr.14201) 1046 : cluster [DBG] pgmap v618: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:47.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:47 vm07 bash[17031]: cluster 2026-03-07T07:13:46.343527+0000 mgr.vm07.yrfcuj (mgr.14201) 1046 : cluster [DBG] pgmap v618: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:47.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:47 vm07 bash[17031]: cluster 2026-03-07T07:13:46.343527+0000 mgr.vm07.yrfcuj (mgr.14201) 1046 : cluster [DBG] pgmap v618: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:48.178 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:13:48.336 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:13:48.336 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (10m) 9m ago 16m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:13:48.336 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (9m) 9m ago 16m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:13:48.336 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 3m ago 16m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:13:48.336 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 3m ago 16m - - 2026-03-07T08:13:48.563 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:13:48.563 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:13:48.563 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:13:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:49 vm09 bash[21220]: audit 2026-03-07T07:13:48.167205+0000 mgr.vm07.yrfcuj (mgr.14201) 1047 : audit [DBG] from='client.16730 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:49 vm09 bash[21220]: audit 2026-03-07T07:13:48.167205+0000 mgr.vm07.yrfcuj (mgr.14201) 1047 : audit [DBG] from='client.16730 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:49 vm09 bash[21220]: audit 2026-03-07T07:13:48.338207+0000 mgr.vm07.yrfcuj (mgr.14201) 1048 : audit [DBG] from='client.16734 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:49 vm09 bash[21220]: audit 2026-03-07T07:13:48.338207+0000 mgr.vm07.yrfcuj (mgr.14201) 1048 : audit [DBG] from='client.16734 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:49 vm09 bash[21220]: cluster 2026-03-07T07:13:48.344054+0000 mgr.vm07.yrfcuj (mgr.14201) 1049 : cluster [DBG] pgmap v619: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:49 vm09 bash[21220]: cluster 2026-03-07T07:13:48.344054+0000 mgr.vm07.yrfcuj (mgr.14201) 1049 : cluster [DBG] pgmap v619: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:49 vm09 bash[21220]: audit 2026-03-07T07:13:48.568168+0000 mon.vm07 (mon.0) 1213 : audit [DBG] from='client.? 192.168.123.107:0/1305288941' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:49 vm09 bash[21220]: audit 2026-03-07T07:13:48.568168+0000 mon.vm07 (mon.0) 1213 : audit [DBG] from='client.? 192.168.123.107:0/1305288941' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:49.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:49 vm07 bash[17031]: audit 2026-03-07T07:13:48.167205+0000 mgr.vm07.yrfcuj (mgr.14201) 1047 : audit [DBG] from='client.16730 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:49.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:49 vm07 bash[17031]: audit 2026-03-07T07:13:48.167205+0000 mgr.vm07.yrfcuj (mgr.14201) 1047 : audit [DBG] from='client.16730 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:49.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:49 vm07 bash[17031]: audit 2026-03-07T07:13:48.338207+0000 mgr.vm07.yrfcuj (mgr.14201) 1048 : audit [DBG] from='client.16734 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:49.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:49 vm07 bash[17031]: audit 2026-03-07T07:13:48.338207+0000 mgr.vm07.yrfcuj (mgr.14201) 1048 : audit [DBG] from='client.16734 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:49.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:49 vm07 bash[17031]: cluster 2026-03-07T07:13:48.344054+0000 mgr.vm07.yrfcuj (mgr.14201) 1049 : cluster [DBG] pgmap v619: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:49.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:49 vm07 bash[17031]: cluster 2026-03-07T07:13:48.344054+0000 mgr.vm07.yrfcuj (mgr.14201) 1049 : cluster [DBG] pgmap v619: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:49.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:49 vm07 bash[17031]: audit 2026-03-07T07:13:48.568168+0000 mon.vm07 (mon.0) 1213 : audit [DBG] from='client.? 192.168.123.107:0/1305288941' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:49.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:49 vm07 bash[17031]: audit 2026-03-07T07:13:48.568168+0000 mon.vm07 (mon.0) 1213 : audit [DBG] from='client.? 192.168.123.107:0/1305288941' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:51.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:51 vm07 bash[17031]: cluster 2026-03-07T07:13:50.344428+0000 mgr.vm07.yrfcuj (mgr.14201) 1050 : cluster [DBG] pgmap v620: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:51.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:51 vm07 bash[17031]: cluster 2026-03-07T07:13:50.344428+0000 mgr.vm07.yrfcuj (mgr.14201) 1050 : cluster [DBG] pgmap v620: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:51 vm09 bash[21220]: cluster 2026-03-07T07:13:50.344428+0000 mgr.vm07.yrfcuj (mgr.14201) 1050 : cluster [DBG] pgmap v620: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:51 vm09 bash[21220]: cluster 2026-03-07T07:13:50.344428+0000 mgr.vm07.yrfcuj (mgr.14201) 1050 : cluster [DBG] pgmap v620: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:53.745 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:13:53.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:53 vm09 bash[21220]: cluster 2026-03-07T07:13:52.344924+0000 mgr.vm07.yrfcuj (mgr.14201) 1051 : cluster [DBG] pgmap v621: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:53.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:53 vm09 bash[21220]: cluster 2026-03-07T07:13:52.344924+0000 mgr.vm07.yrfcuj (mgr.14201) 1051 : cluster [DBG] pgmap v621: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:53.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:53 vm07 bash[17031]: cluster 2026-03-07T07:13:52.344924+0000 mgr.vm07.yrfcuj (mgr.14201) 1051 : cluster [DBG] pgmap v621: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:53.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:53 vm07 bash[17031]: cluster 2026-03-07T07:13:52.344924+0000 mgr.vm07.yrfcuj (mgr.14201) 1051 : cluster [DBG] pgmap v621: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:13:53.916 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:13:53.916 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (10m) 9m ago 16m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:13:53.916 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (9m) 9m ago 16m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:13:53.916 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 4m ago 16m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:13:53.917 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 4m ago 16m - - 2026-03-07T08:13:54.127 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:13:54.127 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:13:54.127 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:13:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:54 vm09 bash[21220]: audit 2026-03-07T07:13:54.132685+0000 mon.vm07 (mon.0) 1214 : audit [DBG] from='client.? 192.168.123.107:0/3349067600' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:54 vm09 bash[21220]: audit 2026-03-07T07:13:54.132685+0000 mon.vm07 (mon.0) 1214 : audit [DBG] from='client.? 192.168.123.107:0/3349067600' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:54 vm09 bash[21220]: audit 2026-03-07T07:13:54.381789+0000 mon.vm07 (mon.0) 1215 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:13:54.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:54 vm09 bash[21220]: audit 2026-03-07T07:13:54.381789+0000 mon.vm07 (mon.0) 1215 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:13:54.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:54 vm07 bash[17031]: audit 2026-03-07T07:13:54.132685+0000 mon.vm07 (mon.0) 1214 : audit [DBG] from='client.? 192.168.123.107:0/3349067600' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:54.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:54 vm07 bash[17031]: audit 2026-03-07T07:13:54.132685+0000 mon.vm07 (mon.0) 1214 : audit [DBG] from='client.? 192.168.123.107:0/3349067600' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:13:54.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:54 vm07 bash[17031]: audit 2026-03-07T07:13:54.381789+0000 mon.vm07 (mon.0) 1215 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:13:54.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:54 vm07 bash[17031]: audit 2026-03-07T07:13:54.381789+0000 mon.vm07 (mon.0) 1215 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: audit 2026-03-07T07:13:53.734586+0000 mgr.vm07.yrfcuj (mgr.14201) 1052 : audit [DBG] from='client.16742 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: audit 2026-03-07T07:13:53.734586+0000 mgr.vm07.yrfcuj (mgr.14201) 1052 : audit [DBG] from='client.16742 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: audit 2026-03-07T07:13:53.918048+0000 mgr.vm07.yrfcuj (mgr.14201) 1053 : audit [DBG] from='client.16746 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: audit 2026-03-07T07:13:53.918048+0000 mgr.vm07.yrfcuj (mgr.14201) 1053 : audit [DBG] from='client.16746 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: cluster 2026-03-07T07:13:54.345403+0000 mgr.vm07.yrfcuj (mgr.14201) 1054 : cluster [DBG] pgmap v622: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: cluster 2026-03-07T07:13:54.345403+0000 mgr.vm07.yrfcuj (mgr.14201) 1054 : cluster [DBG] pgmap v622: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: audit 2026-03-07T07:13:54.671507+0000 mon.vm07 (mon.0) 1216 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: audit 2026-03-07T07:13:54.671507+0000 mon.vm07 (mon.0) 1216 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: audit 2026-03-07T07:13:54.676280+0000 mon.vm07 (mon.0) 1217 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: audit 2026-03-07T07:13:54.676280+0000 mon.vm07 (mon.0) 1217 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: audit 2026-03-07T07:13:54.997199+0000 mon.vm07 (mon.0) 1218 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: audit 2026-03-07T07:13:54.997199+0000 mon.vm07 (mon.0) 1218 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: audit 2026-03-07T07:13:54.998319+0000 mon.vm07 (mon.0) 1219 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: audit 2026-03-07T07:13:54.998319+0000 mon.vm07 (mon.0) 1219 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: audit 2026-03-07T07:13:55.005307+0000 mon.vm07 (mon.0) 1220 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: audit 2026-03-07T07:13:55.005307+0000 mon.vm07 (mon.0) 1220 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: audit 2026-03-07T07:13:55.007441+0000 mon.vm07 (mon.0) 1221 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: audit 2026-03-07T07:13:55.007441+0000 mon.vm07 (mon.0) 1221 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: audit 2026-03-07T07:13:55.145253+0000 mon.vm07 (mon.0) 1222 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:13:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:55 vm09 bash[21220]: audit 2026-03-07T07:13:55.145253+0000 mon.vm07 (mon.0) 1222 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:13:55.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: audit 2026-03-07T07:13:53.734586+0000 mgr.vm07.yrfcuj (mgr.14201) 1052 : audit [DBG] from='client.16742 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:55.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: audit 2026-03-07T07:13:53.734586+0000 mgr.vm07.yrfcuj (mgr.14201) 1052 : audit [DBG] from='client.16742 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:55.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: audit 2026-03-07T07:13:53.918048+0000 mgr.vm07.yrfcuj (mgr.14201) 1053 : audit [DBG] from='client.16746 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:55.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: audit 2026-03-07T07:13:53.918048+0000 mgr.vm07.yrfcuj (mgr.14201) 1053 : audit [DBG] from='client.16746 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:13:55.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: cluster 2026-03-07T07:13:54.345403+0000 mgr.vm07.yrfcuj (mgr.14201) 1054 : cluster [DBG] pgmap v622: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:55.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: cluster 2026-03-07T07:13:54.345403+0000 mgr.vm07.yrfcuj (mgr.14201) 1054 : cluster [DBG] pgmap v622: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:55.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: audit 2026-03-07T07:13:54.671507+0000 mon.vm07 (mon.0) 1216 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:13:55.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: audit 2026-03-07T07:13:54.671507+0000 mon.vm07 (mon.0) 1216 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:13:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: audit 2026-03-07T07:13:54.676280+0000 mon.vm07 (mon.0) 1217 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:13:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: audit 2026-03-07T07:13:54.676280+0000 mon.vm07 (mon.0) 1217 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:13:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: audit 2026-03-07T07:13:54.997199+0000 mon.vm07 (mon.0) 1218 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:13:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: audit 2026-03-07T07:13:54.997199+0000 mon.vm07 (mon.0) 1218 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:13:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: audit 2026-03-07T07:13:54.998319+0000 mon.vm07 (mon.0) 1219 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:13:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: audit 2026-03-07T07:13:54.998319+0000 mon.vm07 (mon.0) 1219 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:13:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: audit 2026-03-07T07:13:55.005307+0000 mon.vm07 (mon.0) 1220 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:13:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: audit 2026-03-07T07:13:55.005307+0000 mon.vm07 (mon.0) 1220 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:13:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: audit 2026-03-07T07:13:55.007441+0000 mon.vm07 (mon.0) 1221 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:13:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: audit 2026-03-07T07:13:55.007441+0000 mon.vm07 (mon.0) 1221 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:13:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: audit 2026-03-07T07:13:55.145253+0000 mon.vm07 (mon.0) 1222 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:13:55.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:55 vm07 bash[17031]: audit 2026-03-07T07:13:55.145253+0000 mon.vm07 (mon.0) 1222 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:13:56.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:56 vm09 bash[21220]: cluster 2026-03-07T07:13:54.999901+0000 mgr.vm07.yrfcuj (mgr.14201) 1055 : cluster [DBG] pgmap v623: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:56.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:56 vm09 bash[21220]: cluster 2026-03-07T07:13:54.999901+0000 mgr.vm07.yrfcuj (mgr.14201) 1055 : cluster [DBG] pgmap v623: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:56.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:56 vm09 bash[21220]: cluster 2026-03-07T07:13:55.000055+0000 mgr.vm07.yrfcuj (mgr.14201) 1056 : cluster [DBG] pgmap v624: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:56.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:56 vm09 bash[21220]: cluster 2026-03-07T07:13:55.000055+0000 mgr.vm07.yrfcuj (mgr.14201) 1056 : cluster [DBG] pgmap v624: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:56.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:56 vm07 bash[17031]: cluster 2026-03-07T07:13:54.999901+0000 mgr.vm07.yrfcuj (mgr.14201) 1055 : cluster [DBG] pgmap v623: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:56.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:56 vm07 bash[17031]: cluster 2026-03-07T07:13:54.999901+0000 mgr.vm07.yrfcuj (mgr.14201) 1055 : cluster [DBG] pgmap v623: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:56.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:56 vm07 bash[17031]: cluster 2026-03-07T07:13:55.000055+0000 mgr.vm07.yrfcuj (mgr.14201) 1056 : cluster [DBG] pgmap v624: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:56.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:56 vm07 bash[17031]: cluster 2026-03-07T07:13:55.000055+0000 mgr.vm07.yrfcuj (mgr.14201) 1056 : cluster [DBG] pgmap v624: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:13:58.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:58 vm09 bash[21220]: cluster 2026-03-07T07:13:57.000474+0000 mgr.vm07.yrfcuj (mgr.14201) 1057 : cluster [DBG] pgmap v625: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 118 B/s rd, 236 B/s wr, 0 op/s 2026-03-07T08:13:58.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:13:58 vm09 bash[21220]: cluster 2026-03-07T07:13:57.000474+0000 mgr.vm07.yrfcuj (mgr.14201) 1057 : cluster [DBG] pgmap v625: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 118 B/s rd, 236 B/s wr, 0 op/s 2026-03-07T08:13:58.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:58 vm07 bash[17031]: cluster 2026-03-07T07:13:57.000474+0000 mgr.vm07.yrfcuj (mgr.14201) 1057 : cluster [DBG] pgmap v625: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 118 B/s rd, 236 B/s wr, 0 op/s 2026-03-07T08:13:58.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:13:58 vm07 bash[17031]: cluster 2026-03-07T07:13:57.000474+0000 mgr.vm07.yrfcuj (mgr.14201) 1057 : cluster [DBG] pgmap v625: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 118 B/s rd, 236 B/s wr, 0 op/s 2026-03-07T08:13:59.318 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:13:59.472 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:13:59.473 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (10m) 9m ago 16m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:13:59.473 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (9m) 9m ago 16m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:13:59.473 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 4m ago 16m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:13:59.473 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 4m ago 16m - - 2026-03-07T08:13:59.671 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:13:59.671 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:13:59.671 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:14:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:00 vm09 bash[21220]: cluster 2026-03-07T07:13:59.000943+0000 mgr.vm07.yrfcuj (mgr.14201) 1058 : cluster [DBG] pgmap v626: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 118 B/s rd, 236 B/s wr, 0 op/s 2026-03-07T08:14:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:00 vm09 bash[21220]: cluster 2026-03-07T07:13:59.000943+0000 mgr.vm07.yrfcuj (mgr.14201) 1058 : cluster [DBG] pgmap v626: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 118 B/s rd, 236 B/s wr, 0 op/s 2026-03-07T08:14:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:00 vm09 bash[21220]: audit 2026-03-07T07:13:59.308458+0000 mgr.vm07.yrfcuj (mgr.14201) 1059 : audit [DBG] from='client.25791 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:00 vm09 bash[21220]: audit 2026-03-07T07:13:59.308458+0000 mgr.vm07.yrfcuj (mgr.14201) 1059 : audit [DBG] from='client.25791 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:00 vm09 bash[21220]: audit 2026-03-07T07:13:59.475329+0000 mgr.vm07.yrfcuj (mgr.14201) 1060 : audit [DBG] from='client.16758 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:00 vm09 bash[21220]: audit 2026-03-07T07:13:59.475329+0000 mgr.vm07.yrfcuj (mgr.14201) 1060 : audit [DBG] from='client.16758 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:00 vm09 bash[21220]: audit 2026-03-07T07:13:59.676317+0000 mon.vm07 (mon.0) 1223 : audit [DBG] from='client.? 192.168.123.107:0/3238826013' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:00.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:00 vm09 bash[21220]: audit 2026-03-07T07:13:59.676317+0000 mon.vm07 (mon.0) 1223 : audit [DBG] from='client.? 192.168.123.107:0/3238826013' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:00.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:00 vm07 bash[17031]: cluster 2026-03-07T07:13:59.000943+0000 mgr.vm07.yrfcuj (mgr.14201) 1058 : cluster [DBG] pgmap v626: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 118 B/s rd, 236 B/s wr, 0 op/s 2026-03-07T08:14:00.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:00 vm07 bash[17031]: cluster 2026-03-07T07:13:59.000943+0000 mgr.vm07.yrfcuj (mgr.14201) 1058 : cluster [DBG] pgmap v626: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 118 B/s rd, 236 B/s wr, 0 op/s 2026-03-07T08:14:00.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:00 vm07 bash[17031]: audit 2026-03-07T07:13:59.308458+0000 mgr.vm07.yrfcuj (mgr.14201) 1059 : audit [DBG] from='client.25791 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:00.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:00 vm07 bash[17031]: audit 2026-03-07T07:13:59.308458+0000 mgr.vm07.yrfcuj (mgr.14201) 1059 : audit [DBG] from='client.25791 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:00.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:00 vm07 bash[17031]: audit 2026-03-07T07:13:59.475329+0000 mgr.vm07.yrfcuj (mgr.14201) 1060 : audit [DBG] from='client.16758 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:00.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:00 vm07 bash[17031]: audit 2026-03-07T07:13:59.475329+0000 mgr.vm07.yrfcuj (mgr.14201) 1060 : audit [DBG] from='client.16758 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:00.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:00 vm07 bash[17031]: audit 2026-03-07T07:13:59.676317+0000 mon.vm07 (mon.0) 1223 : audit [DBG] from='client.? 192.168.123.107:0/3238826013' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:00.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:00 vm07 bash[17031]: audit 2026-03-07T07:13:59.676317+0000 mon.vm07 (mon.0) 1223 : audit [DBG] from='client.? 192.168.123.107:0/3238826013' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:02.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:02 vm09 bash[21220]: cluster 2026-03-07T07:14:01.001420+0000 mgr.vm07.yrfcuj (mgr.14201) 1061 : cluster [DBG] pgmap v627: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 118 B/s rd, 236 B/s wr, 0 op/s 2026-03-07T08:14:02.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:02 vm09 bash[21220]: cluster 2026-03-07T07:14:01.001420+0000 mgr.vm07.yrfcuj (mgr.14201) 1061 : cluster [DBG] pgmap v627: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 118 B/s rd, 236 B/s wr, 0 op/s 2026-03-07T08:14:02.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:02 vm07 bash[17031]: cluster 2026-03-07T07:14:01.001420+0000 mgr.vm07.yrfcuj (mgr.14201) 1061 : cluster [DBG] pgmap v627: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 118 B/s rd, 236 B/s wr, 0 op/s 2026-03-07T08:14:02.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:02 vm07 bash[17031]: cluster 2026-03-07T07:14:01.001420+0000 mgr.vm07.yrfcuj (mgr.14201) 1061 : cluster [DBG] pgmap v627: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 118 B/s rd, 236 B/s wr, 0 op/s 2026-03-07T08:14:04.859 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:14:04.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:04 vm09 bash[21220]: cluster 2026-03-07T07:14:03.001933+0000 mgr.vm07.yrfcuj (mgr.14201) 1062 : cluster [DBG] pgmap v628: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 236 B/s rd, 473 B/s wr, 0 op/s 2026-03-07T08:14:04.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:04 vm09 bash[21220]: cluster 2026-03-07T07:14:03.001933+0000 mgr.vm07.yrfcuj (mgr.14201) 1062 : cluster [DBG] pgmap v628: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 236 B/s rd, 473 B/s wr, 0 op/s 2026-03-07T08:14:04.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:04 vm07 bash[17031]: cluster 2026-03-07T07:14:03.001933+0000 mgr.vm07.yrfcuj (mgr.14201) 1062 : cluster [DBG] pgmap v628: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 236 B/s rd, 473 B/s wr, 0 op/s 2026-03-07T08:14:04.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:04 vm07 bash[17031]: cluster 2026-03-07T07:14:03.001933+0000 mgr.vm07.yrfcuj (mgr.14201) 1062 : cluster [DBG] pgmap v628: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 236 B/s rd, 473 B/s wr, 0 op/s 2026-03-07T08:14:05.013 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:14:05.013 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (10m) 9m ago 16m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:14:05.013 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (9m) 9m ago 16m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:14:05.013 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 4m ago 16m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:14:05.013 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 4m ago 16m - - 2026-03-07T08:14:05.224 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:14:05.224 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:14:05.224 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:14:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:05 vm09 bash[21220]: audit 2026-03-07T07:14:05.229702+0000 mon.vm07 (mon.0) 1224 : audit [DBG] from='client.? 192.168.123.107:0/2625699273' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:05 vm09 bash[21220]: audit 2026-03-07T07:14:05.229702+0000 mon.vm07 (mon.0) 1224 : audit [DBG] from='client.? 192.168.123.107:0/2625699273' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:05.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:05 vm07 bash[17031]: audit 2026-03-07T07:14:05.229702+0000 mon.vm07 (mon.0) 1224 : audit [DBG] from='client.? 192.168.123.107:0/2625699273' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:05.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:05 vm07 bash[17031]: audit 2026-03-07T07:14:05.229702+0000 mon.vm07 (mon.0) 1224 : audit [DBG] from='client.? 192.168.123.107:0/2625699273' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:06 vm09 bash[21220]: audit 2026-03-07T07:14:04.848228+0000 mgr.vm07.yrfcuj (mgr.14201) 1063 : audit [DBG] from='client.16766 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:06 vm09 bash[21220]: audit 2026-03-07T07:14:04.848228+0000 mgr.vm07.yrfcuj (mgr.14201) 1063 : audit [DBG] from='client.16766 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:06 vm09 bash[21220]: cluster 2026-03-07T07:14:05.002470+0000 mgr.vm07.yrfcuj (mgr.14201) 1064 : cluster [DBG] pgmap v629: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 204 B/s rd, 409 B/s wr, 0 op/s 2026-03-07T08:14:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:06 vm09 bash[21220]: cluster 2026-03-07T07:14:05.002470+0000 mgr.vm07.yrfcuj (mgr.14201) 1064 : cluster [DBG] pgmap v629: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 204 B/s rd, 409 B/s wr, 0 op/s 2026-03-07T08:14:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:06 vm09 bash[21220]: audit 2026-03-07T07:14:05.014621+0000 mgr.vm07.yrfcuj (mgr.14201) 1065 : audit [DBG] from='client.16770 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:06.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:06 vm09 bash[21220]: audit 2026-03-07T07:14:05.014621+0000 mgr.vm07.yrfcuj (mgr.14201) 1065 : audit [DBG] from='client.16770 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:06.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:06 vm07 bash[17031]: audit 2026-03-07T07:14:04.848228+0000 mgr.vm07.yrfcuj (mgr.14201) 1063 : audit [DBG] from='client.16766 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:06.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:06 vm07 bash[17031]: audit 2026-03-07T07:14:04.848228+0000 mgr.vm07.yrfcuj (mgr.14201) 1063 : audit [DBG] from='client.16766 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:06.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:06 vm07 bash[17031]: cluster 2026-03-07T07:14:05.002470+0000 mgr.vm07.yrfcuj (mgr.14201) 1064 : cluster [DBG] pgmap v629: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 204 B/s rd, 409 B/s wr, 0 op/s 2026-03-07T08:14:06.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:06 vm07 bash[17031]: cluster 2026-03-07T07:14:05.002470+0000 mgr.vm07.yrfcuj (mgr.14201) 1064 : cluster [DBG] pgmap v629: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 204 B/s rd, 409 B/s wr, 0 op/s 2026-03-07T08:14:06.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:06 vm07 bash[17031]: audit 2026-03-07T07:14:05.014621+0000 mgr.vm07.yrfcuj (mgr.14201) 1065 : audit [DBG] from='client.16770 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:06.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:06 vm07 bash[17031]: audit 2026-03-07T07:14:05.014621+0000 mgr.vm07.yrfcuj (mgr.14201) 1065 : audit [DBG] from='client.16770 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:08.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:08 vm09 bash[21220]: cluster 2026-03-07T07:14:07.002998+0000 mgr.vm07.yrfcuj (mgr.14201) 1066 : cluster [DBG] pgmap v630: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:14:08.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:08 vm09 bash[21220]: cluster 2026-03-07T07:14:07.002998+0000 mgr.vm07.yrfcuj (mgr.14201) 1066 : cluster [DBG] pgmap v630: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:14:08.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:08 vm07 bash[17031]: cluster 2026-03-07T07:14:07.002998+0000 mgr.vm07.yrfcuj (mgr.14201) 1066 : cluster [DBG] pgmap v630: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:14:08.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:08 vm07 bash[17031]: cluster 2026-03-07T07:14:07.002998+0000 mgr.vm07.yrfcuj (mgr.14201) 1066 : cluster [DBG] pgmap v630: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:14:10.413 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:14:10.585 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:14:10.585 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (10m) 9m ago 16m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:14:10.585 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (10m) 9m ago 16m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:14:10.585 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 4m ago 16m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:14:10.585 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 4m ago 16m - - 2026-03-07T08:14:10.777 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:14:10.777 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:14:10.777 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:14:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:10 vm09 bash[21220]: cluster 2026-03-07T07:14:09.003476+0000 mgr.vm07.yrfcuj (mgr.14201) 1067 : cluster [DBG] pgmap v631: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 3.6 KiB/s rd, 170 B/s wr, 6 op/s 2026-03-07T08:14:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:10 vm09 bash[21220]: cluster 2026-03-07T07:14:09.003476+0000 mgr.vm07.yrfcuj (mgr.14201) 1067 : cluster [DBG] pgmap v631: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 3.6 KiB/s rd, 170 B/s wr, 6 op/s 2026-03-07T08:14:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:10 vm09 bash[21220]: audit 2026-03-07T07:14:10.145887+0000 mon.vm07 (mon.0) 1225 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:14:10.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:10 vm09 bash[21220]: audit 2026-03-07T07:14:10.145887+0000 mon.vm07 (mon.0) 1225 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:14:10.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:10 vm07 bash[17031]: cluster 2026-03-07T07:14:09.003476+0000 mgr.vm07.yrfcuj (mgr.14201) 1067 : cluster [DBG] pgmap v631: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 3.6 KiB/s rd, 170 B/s wr, 6 op/s 2026-03-07T08:14:10.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:10 vm07 bash[17031]: cluster 2026-03-07T07:14:09.003476+0000 mgr.vm07.yrfcuj (mgr.14201) 1067 : cluster [DBG] pgmap v631: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 3.6 KiB/s rd, 170 B/s wr, 6 op/s 2026-03-07T08:14:10.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:10 vm07 bash[17031]: audit 2026-03-07T07:14:10.145887+0000 mon.vm07 (mon.0) 1225 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:14:10.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:10 vm07 bash[17031]: audit 2026-03-07T07:14:10.145887+0000 mon.vm07 (mon.0) 1225 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:14:11.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:11 vm09 bash[21220]: audit 2026-03-07T07:14:10.403049+0000 mgr.vm07.yrfcuj (mgr.14201) 1068 : audit [DBG] from='client.16778 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:11.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:11 vm09 bash[21220]: audit 2026-03-07T07:14:10.403049+0000 mgr.vm07.yrfcuj (mgr.14201) 1068 : audit [DBG] from='client.16778 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:11.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:11 vm09 bash[21220]: audit 2026-03-07T07:14:10.586114+0000 mgr.vm07.yrfcuj (mgr.14201) 1069 : audit [DBG] from='client.16782 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:11.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:11 vm09 bash[21220]: audit 2026-03-07T07:14:10.586114+0000 mgr.vm07.yrfcuj (mgr.14201) 1069 : audit [DBG] from='client.16782 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:11.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:11 vm09 bash[21220]: audit 2026-03-07T07:14:10.782791+0000 mon.vm07 (mon.0) 1226 : audit [DBG] from='client.? 192.168.123.107:0/3325085078' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:11.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:11 vm09 bash[21220]: audit 2026-03-07T07:14:10.782791+0000 mon.vm07 (mon.0) 1226 : audit [DBG] from='client.? 192.168.123.107:0/3325085078' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:11.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:11 vm07 bash[17031]: audit 2026-03-07T07:14:10.403049+0000 mgr.vm07.yrfcuj (mgr.14201) 1068 : audit [DBG] from='client.16778 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:11.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:11 vm07 bash[17031]: audit 2026-03-07T07:14:10.403049+0000 mgr.vm07.yrfcuj (mgr.14201) 1068 : audit [DBG] from='client.16778 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:11.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:11 vm07 bash[17031]: audit 2026-03-07T07:14:10.586114+0000 mgr.vm07.yrfcuj (mgr.14201) 1069 : audit [DBG] from='client.16782 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:11.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:11 vm07 bash[17031]: audit 2026-03-07T07:14:10.586114+0000 mgr.vm07.yrfcuj (mgr.14201) 1069 : audit [DBG] from='client.16782 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:11.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:11 vm07 bash[17031]: audit 2026-03-07T07:14:10.782791+0000 mon.vm07 (mon.0) 1226 : audit [DBG] from='client.? 192.168.123.107:0/3325085078' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:11.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:11 vm07 bash[17031]: audit 2026-03-07T07:14:10.782791+0000 mon.vm07 (mon.0) 1226 : audit [DBG] from='client.? 192.168.123.107:0/3325085078' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:12.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:12 vm09 bash[21220]: cluster 2026-03-07T07:14:11.003950+0000 mgr.vm07.yrfcuj (mgr.14201) 1070 : cluster [DBG] pgmap v632: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 8.8 KiB/s rd, 170 B/s wr, 14 op/s 2026-03-07T08:14:12.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:12 vm09 bash[21220]: cluster 2026-03-07T07:14:11.003950+0000 mgr.vm07.yrfcuj (mgr.14201) 1070 : cluster [DBG] pgmap v632: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 8.8 KiB/s rd, 170 B/s wr, 14 op/s 2026-03-07T08:14:12.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:12 vm07 bash[17031]: cluster 2026-03-07T07:14:11.003950+0000 mgr.vm07.yrfcuj (mgr.14201) 1070 : cluster [DBG] pgmap v632: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 8.8 KiB/s rd, 170 B/s wr, 14 op/s 2026-03-07T08:14:12.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:12 vm07 bash[17031]: cluster 2026-03-07T07:14:11.003950+0000 mgr.vm07.yrfcuj (mgr.14201) 1070 : cluster [DBG] pgmap v632: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 8.8 KiB/s rd, 170 B/s wr, 14 op/s 2026-03-07T08:14:14.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:14 vm09 bash[21220]: cluster 2026-03-07T07:14:13.004514+0000 mgr.vm07.yrfcuj (mgr.14201) 1071 : cluster [DBG] pgmap v633: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:14:14.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:14 vm09 bash[21220]: cluster 2026-03-07T07:14:13.004514+0000 mgr.vm07.yrfcuj (mgr.14201) 1071 : cluster [DBG] pgmap v633: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:14:14.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:14 vm07 bash[17031]: cluster 2026-03-07T07:14:13.004514+0000 mgr.vm07.yrfcuj (mgr.14201) 1071 : cluster [DBG] pgmap v633: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:14:14.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:14 vm07 bash[17031]: cluster 2026-03-07T07:14:13.004514+0000 mgr.vm07.yrfcuj (mgr.14201) 1071 : cluster [DBG] pgmap v633: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 341 B/s wr, 60 op/s 2026-03-07T08:14:15.955 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:14:16.266 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:14:16.266 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (10m) 10m ago 16m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:14:16.266 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (10m) 10m ago 16m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:14:16.266 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 4m ago 16m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:14:16.266 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 4m ago 16m - - 2026-03-07T08:14:16.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:16 vm09 bash[21220]: cluster 2026-03-07T07:14:15.005040+0000 mgr.vm07.yrfcuj (mgr.14201) 1072 : cluster [DBG] pgmap v634: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:14:16.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:16 vm09 bash[21220]: cluster 2026-03-07T07:14:15.005040+0000 mgr.vm07.yrfcuj (mgr.14201) 1072 : cluster [DBG] pgmap v634: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:14:16.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:15 vm07 bash[17031]: cluster 2026-03-07T07:14:15.005040+0000 mgr.vm07.yrfcuj (mgr.14201) 1072 : cluster [DBG] pgmap v634: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:14:16.398 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:15 vm07 bash[17031]: cluster 2026-03-07T07:14:15.005040+0000 mgr.vm07.yrfcuj (mgr.14201) 1072 : cluster [DBG] pgmap v634: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:14:16.469 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:14:16.470 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:14:16.470 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:14:17.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:17 vm09 bash[21220]: audit 2026-03-07T07:14:15.943381+0000 mgr.vm07.yrfcuj (mgr.14201) 1073 : audit [DBG] from='client.16790 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:17.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:17 vm09 bash[21220]: audit 2026-03-07T07:14:15.943381+0000 mgr.vm07.yrfcuj (mgr.14201) 1073 : audit [DBG] from='client.16790 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:17.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:17 vm09 bash[21220]: audit 2026-03-07T07:14:16.268738+0000 mgr.vm07.yrfcuj (mgr.14201) 1074 : audit [DBG] from='client.16794 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:17.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:17 vm09 bash[21220]: audit 2026-03-07T07:14:16.268738+0000 mgr.vm07.yrfcuj (mgr.14201) 1074 : audit [DBG] from='client.16794 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:17.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:17 vm09 bash[21220]: audit 2026-03-07T07:14:16.474931+0000 mon.vm07 (mon.0) 1227 : audit [DBG] from='client.? 192.168.123.107:0/3257604757' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:17.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:17 vm09 bash[21220]: audit 2026-03-07T07:14:16.474931+0000 mon.vm07 (mon.0) 1227 : audit [DBG] from='client.? 192.168.123.107:0/3257604757' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:17.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:16 vm07 bash[17031]: audit 2026-03-07T07:14:15.943381+0000 mgr.vm07.yrfcuj (mgr.14201) 1073 : audit [DBG] from='client.16790 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:17.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:16 vm07 bash[17031]: audit 2026-03-07T07:14:15.943381+0000 mgr.vm07.yrfcuj (mgr.14201) 1073 : audit [DBG] from='client.16790 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:17.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:16 vm07 bash[17031]: audit 2026-03-07T07:14:16.268738+0000 mgr.vm07.yrfcuj (mgr.14201) 1074 : audit [DBG] from='client.16794 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:17.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:16 vm07 bash[17031]: audit 2026-03-07T07:14:16.268738+0000 mgr.vm07.yrfcuj (mgr.14201) 1074 : audit [DBG] from='client.16794 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:17.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:16 vm07 bash[17031]: audit 2026-03-07T07:14:16.474931+0000 mon.vm07 (mon.0) 1227 : audit [DBG] from='client.? 192.168.123.107:0/3257604757' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:17.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:16 vm07 bash[17031]: audit 2026-03-07T07:14:16.474931+0000 mon.vm07 (mon.0) 1227 : audit [DBG] from='client.? 192.168.123.107:0/3257604757' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:18 vm09 bash[21220]: cluster 2026-03-07T07:14:17.005498+0000 mgr.vm07.yrfcuj (mgr.14201) 1075 : cluster [DBG] pgmap v635: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:14:18.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:18 vm09 bash[21220]: cluster 2026-03-07T07:14:17.005498+0000 mgr.vm07.yrfcuj (mgr.14201) 1075 : cluster [DBG] pgmap v635: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:14:18.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:18 vm07 bash[17031]: cluster 2026-03-07T07:14:17.005498+0000 mgr.vm07.yrfcuj (mgr.14201) 1075 : cluster [DBG] pgmap v635: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:14:18.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:18 vm07 bash[17031]: cluster 2026-03-07T07:14:17.005498+0000 mgr.vm07.yrfcuj (mgr.14201) 1075 : cluster [DBG] pgmap v635: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:14:20.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:20 vm09 bash[21220]: cluster 2026-03-07T07:14:19.006073+0000 mgr.vm07.yrfcuj (mgr.14201) 1076 : cluster [DBG] pgmap v636: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:14:20.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:20 vm09 bash[21220]: cluster 2026-03-07T07:14:19.006073+0000 mgr.vm07.yrfcuj (mgr.14201) 1076 : cluster [DBG] pgmap v636: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:14:20.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:20 vm07 bash[17031]: cluster 2026-03-07T07:14:19.006073+0000 mgr.vm07.yrfcuj (mgr.14201) 1076 : cluster [DBG] pgmap v636: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:14:20.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:20 vm07 bash[17031]: cluster 2026-03-07T07:14:19.006073+0000 mgr.vm07.yrfcuj (mgr.14201) 1076 : cluster [DBG] pgmap v636: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 170 B/s wr, 59 op/s 2026-03-07T08:14:21.660 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:14:21.824 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:14:21.824 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (10m) 10m ago 16m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:14:21.824 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (10m) 10m ago 16m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:14:21.824 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (4m) 4m ago 16m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:14:21.824 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 4m ago 16m - - 2026-03-07T08:14:22.035 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:14:22.035 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:14:22.035 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:14:22.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:22 vm09 bash[21220]: cluster 2026-03-07T07:14:21.006490+0000 mgr.vm07.yrfcuj (mgr.14201) 1077 : cluster [DBG] pgmap v637: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 32 KiB/s rd, 170 B/s wr, 53 op/s 2026-03-07T08:14:22.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:22 vm09 bash[21220]: cluster 2026-03-07T07:14:21.006490+0000 mgr.vm07.yrfcuj (mgr.14201) 1077 : cluster [DBG] pgmap v637: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 32 KiB/s rd, 170 B/s wr, 53 op/s 2026-03-07T08:14:22.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:22 vm09 bash[21220]: audit 2026-03-07T07:14:22.040309+0000 mon.vm07 (mon.0) 1228 : audit [DBG] from='client.? 192.168.123.107:0/3663904321' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:22.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:22 vm09 bash[21220]: audit 2026-03-07T07:14:22.040309+0000 mon.vm07 (mon.0) 1228 : audit [DBG] from='client.? 192.168.123.107:0/3663904321' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:22.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:22 vm07 bash[17031]: cluster 2026-03-07T07:14:21.006490+0000 mgr.vm07.yrfcuj (mgr.14201) 1077 : cluster [DBG] pgmap v637: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 32 KiB/s rd, 170 B/s wr, 53 op/s 2026-03-07T08:14:22.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:22 vm07 bash[17031]: cluster 2026-03-07T07:14:21.006490+0000 mgr.vm07.yrfcuj (mgr.14201) 1077 : cluster [DBG] pgmap v637: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 32 KiB/s rd, 170 B/s wr, 53 op/s 2026-03-07T08:14:22.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:22 vm07 bash[17031]: audit 2026-03-07T07:14:22.040309+0000 mon.vm07 (mon.0) 1228 : audit [DBG] from='client.? 192.168.123.107:0/3663904321' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:22.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:22 vm07 bash[17031]: audit 2026-03-07T07:14:22.040309+0000 mon.vm07 (mon.0) 1228 : audit [DBG] from='client.? 192.168.123.107:0/3663904321' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:23.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:23 vm09 bash[21220]: audit 2026-03-07T07:14:21.647524+0000 mgr.vm07.yrfcuj (mgr.14201) 1078 : audit [DBG] from='client.16802 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:23.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:23 vm09 bash[21220]: audit 2026-03-07T07:14:21.647524+0000 mgr.vm07.yrfcuj (mgr.14201) 1078 : audit [DBG] from='client.16802 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:23.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:23 vm09 bash[21220]: audit 2026-03-07T07:14:21.826847+0000 mgr.vm07.yrfcuj (mgr.14201) 1079 : audit [DBG] from='client.16806 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:23.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:23 vm09 bash[21220]: audit 2026-03-07T07:14:21.826847+0000 mgr.vm07.yrfcuj (mgr.14201) 1079 : audit [DBG] from='client.16806 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:23.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:23 vm07 bash[17031]: audit 2026-03-07T07:14:21.647524+0000 mgr.vm07.yrfcuj (mgr.14201) 1078 : audit [DBG] from='client.16802 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:23.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:23 vm07 bash[17031]: audit 2026-03-07T07:14:21.647524+0000 mgr.vm07.yrfcuj (mgr.14201) 1078 : audit [DBG] from='client.16802 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:23.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:23 vm07 bash[17031]: audit 2026-03-07T07:14:21.826847+0000 mgr.vm07.yrfcuj (mgr.14201) 1079 : audit [DBG] from='client.16806 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:23.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:23 vm07 bash[17031]: audit 2026-03-07T07:14:21.826847+0000 mgr.vm07.yrfcuj (mgr.14201) 1079 : audit [DBG] from='client.16806 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:24 vm09 bash[21220]: cluster 2026-03-07T07:14:23.007016+0000 mgr.vm07.yrfcuj (mgr.14201) 1080 : cluster [DBG] pgmap v638: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s rd, 170 B/s wr, 45 op/s 2026-03-07T08:14:24.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:24 vm09 bash[21220]: cluster 2026-03-07T07:14:23.007016+0000 mgr.vm07.yrfcuj (mgr.14201) 1080 : cluster [DBG] pgmap v638: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s rd, 170 B/s wr, 45 op/s 2026-03-07T08:14:24.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:24 vm07 bash[17031]: cluster 2026-03-07T07:14:23.007016+0000 mgr.vm07.yrfcuj (mgr.14201) 1080 : cluster [DBG] pgmap v638: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s rd, 170 B/s wr, 45 op/s 2026-03-07T08:14:24.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:24 vm07 bash[17031]: cluster 2026-03-07T07:14:23.007016+0000 mgr.vm07.yrfcuj (mgr.14201) 1080 : cluster [DBG] pgmap v638: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s rd, 170 B/s wr, 45 op/s 2026-03-07T08:14:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:25 vm09 bash[21220]: audit 2026-03-07T07:14:25.145931+0000 mon.vm07 (mon.0) 1229 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:14:25.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:25 vm09 bash[21220]: audit 2026-03-07T07:14:25.145931+0000 mon.vm07 (mon.0) 1229 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:14:25.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:25 vm07 bash[17031]: audit 2026-03-07T07:14:25.145931+0000 mon.vm07 (mon.0) 1229 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:14:25.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:25 vm07 bash[17031]: audit 2026-03-07T07:14:25.145931+0000 mon.vm07 (mon.0) 1229 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:14:26.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:26 vm09 bash[21220]: cluster 2026-03-07T07:14:25.007470+0000 mgr.vm07.yrfcuj (mgr.14201) 1081 : cluster [DBG] pgmap v639: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:14:26.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:26 vm09 bash[21220]: cluster 2026-03-07T07:14:25.007470+0000 mgr.vm07.yrfcuj (mgr.14201) 1081 : cluster [DBG] pgmap v639: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:14:26.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:26 vm07 bash[17031]: cluster 2026-03-07T07:14:25.007470+0000 mgr.vm07.yrfcuj (mgr.14201) 1081 : cluster [DBG] pgmap v639: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:14:26.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:26 vm07 bash[17031]: cluster 2026-03-07T07:14:25.007470+0000 mgr.vm07.yrfcuj (mgr.14201) 1081 : cluster [DBG] pgmap v639: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail 2026-03-07T08:14:27.221 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to stop 2026-03-07T08:14:27.387 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:14:27.387 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (10m) 10m ago 16m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:14:27.387 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (10m) 10m ago 16m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:14:27.387 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (5m) 4m ago 16m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:14:27.387 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 4m ago 16m - - 2026-03-07T08:14:27.584 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:14:27.584 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:14:27.584 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:14:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:28 vm09 bash[21220]: cluster 2026-03-07T07:14:27.008052+0000 mgr.vm07.yrfcuj (mgr.14201) 1082 : cluster [DBG] pgmap v640: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:14:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:28 vm09 bash[21220]: cluster 2026-03-07T07:14:27.008052+0000 mgr.vm07.yrfcuj (mgr.14201) 1082 : cluster [DBG] pgmap v640: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:14:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:28 vm09 bash[21220]: audit 2026-03-07T07:14:27.209875+0000 mgr.vm07.yrfcuj (mgr.14201) 1083 : audit [DBG] from='client.16814 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:28 vm09 bash[21220]: audit 2026-03-07T07:14:27.209875+0000 mgr.vm07.yrfcuj (mgr.14201) 1083 : audit [DBG] from='client.16814 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:28 vm09 bash[21220]: audit 2026-03-07T07:14:27.390167+0000 mgr.vm07.yrfcuj (mgr.14201) 1084 : audit [DBG] from='client.16818 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:28 vm09 bash[21220]: audit 2026-03-07T07:14:27.390167+0000 mgr.vm07.yrfcuj (mgr.14201) 1084 : audit [DBG] from='client.16818 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:28 vm09 bash[21220]: audit 2026-03-07T07:14:27.589665+0000 mon.vm07 (mon.0) 1230 : audit [DBG] from='client.? 192.168.123.107:0/1664252022' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:28.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:28 vm09 bash[21220]: audit 2026-03-07T07:14:27.589665+0000 mon.vm07 (mon.0) 1230 : audit [DBG] from='client.? 192.168.123.107:0/1664252022' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:28.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:28 vm07 bash[17031]: cluster 2026-03-07T07:14:27.008052+0000 mgr.vm07.yrfcuj (mgr.14201) 1082 : cluster [DBG] pgmap v640: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:14:28.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:28 vm07 bash[17031]: cluster 2026-03-07T07:14:27.008052+0000 mgr.vm07.yrfcuj (mgr.14201) 1082 : cluster [DBG] pgmap v640: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:14:28.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:28 vm07 bash[17031]: audit 2026-03-07T07:14:27.209875+0000 mgr.vm07.yrfcuj (mgr.14201) 1083 : audit [DBG] from='client.16814 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:28.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:28 vm07 bash[17031]: audit 2026-03-07T07:14:27.209875+0000 mgr.vm07.yrfcuj (mgr.14201) 1083 : audit [DBG] from='client.16814 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:28.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:28 vm07 bash[17031]: audit 2026-03-07T07:14:27.390167+0000 mgr.vm07.yrfcuj (mgr.14201) 1084 : audit [DBG] from='client.16818 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:28.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:28 vm07 bash[17031]: audit 2026-03-07T07:14:27.390167+0000 mgr.vm07.yrfcuj (mgr.14201) 1084 : audit [DBG] from='client.16818 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:28.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:28 vm07 bash[17031]: audit 2026-03-07T07:14:27.589665+0000 mon.vm07 (mon.0) 1230 : audit [DBG] from='client.? 192.168.123.107:0/1664252022' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:28.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:28 vm07 bash[17031]: audit 2026-03-07T07:14:27.589665+0000 mon.vm07 (mon.0) 1230 : audit [DBG] from='client.? 192.168.123.107:0/1664252022' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:30.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:30 vm09 bash[21220]: cluster 2026-03-07T07:14:29.008561+0000 mgr.vm07.yrfcuj (mgr.14201) 1085 : cluster [DBG] pgmap v641: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:14:30.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:30 vm09 bash[21220]: cluster 2026-03-07T07:14:29.008561+0000 mgr.vm07.yrfcuj (mgr.14201) 1085 : cluster [DBG] pgmap v641: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:14:30.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:30 vm07 bash[17031]: cluster 2026-03-07T07:14:29.008561+0000 mgr.vm07.yrfcuj (mgr.14201) 1085 : cluster [DBG] pgmap v641: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:14:30.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:30 vm07 bash[17031]: cluster 2026-03-07T07:14:29.008561+0000 mgr.vm07.yrfcuj (mgr.14201) 1085 : cluster [DBG] pgmap v641: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:14:32.533 INFO:teuthology.orchestra.run.vm07.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-07T08:14:32.534 INFO:teuthology.orchestra.run.vm07.stderr: Dload Upload Total Spent Left Speed 2026-03-07T08:14:32.534 INFO:teuthology.orchestra.run.vm07.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 182k 0 --:--:-- --:--:-- --:--:-- 182k 2026-03-07T08:14:32.702 INFO:teuthology.orchestra.run.vm07.stdout:anonymousScheduled to start rgw.foo.vm09.kpgoql on host 'vm09' 2026-03-07T08:14:32.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:32 vm09 bash[21220]: cluster 2026-03-07T07:14:31.008918+0000 mgr.vm07.yrfcuj (mgr.14201) 1086 : cluster [DBG] pgmap v642: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:14:32.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:32 vm09 bash[21220]: cluster 2026-03-07T07:14:31.008918+0000 mgr.vm07.yrfcuj (mgr.14201) 1086 : cluster [DBG] pgmap v642: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:14:32.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:32 vm07 bash[17031]: cluster 2026-03-07T07:14:31.008918+0000 mgr.vm07.yrfcuj (mgr.14201) 1086 : cluster [DBG] pgmap v642: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:14:32.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:32 vm07 bash[17031]: cluster 2026-03-07T07:14:31.008918+0000 mgr.vm07.yrfcuj (mgr.14201) 1086 : cluster [DBG] pgmap v642: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:14:32.903 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to start 2026-03-07T08:14:33.079 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:14:33.079 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (11m) 10m ago 16m 94.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:14:33.079 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (10m) 10m ago 16m 89.3M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:14:33.079 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (5m) 4m ago 16m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:14:33.079 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 4m ago 16m - - 2026-03-07T08:14:33.274 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:14:33.274 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:14:33.274 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: audit 2026-03-07T07:14:32.695137+0000 mgr.vm07.yrfcuj (mgr.14201) 1087 : audit [DBG] from='client.16826 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "rgw.foo.vm09.kpgoql", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: audit 2026-03-07T07:14:32.695137+0000 mgr.vm07.yrfcuj (mgr.14201) 1087 : audit [DBG] from='client.16826 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "rgw.foo.vm09.kpgoql", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: cephadm 2026-03-07T07:14:32.695505+0000 mgr.vm07.yrfcuj (mgr.14201) 1088 : cephadm [INF] Schedule start daemon rgw.foo.vm09.kpgoql 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: cephadm 2026-03-07T07:14:32.695505+0000 mgr.vm07.yrfcuj (mgr.14201) 1088 : cephadm [INF] Schedule start daemon rgw.foo.vm09.kpgoql 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: audit 2026-03-07T07:14:32.700760+0000 mon.vm07 (mon.0) 1231 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: audit 2026-03-07T07:14:32.700760+0000 mon.vm07 (mon.0) 1231 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: audit 2026-03-07T07:14:32.706398+0000 mon.vm07 (mon.0) 1232 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: audit 2026-03-07T07:14:32.706398+0000 mon.vm07 (mon.0) 1232 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: audit 2026-03-07T07:14:32.707397+0000 mon.vm07 (mon.0) 1233 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: audit 2026-03-07T07:14:32.707397+0000 mon.vm07 (mon.0) 1233 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: audit 2026-03-07T07:14:32.891319+0000 mgr.vm07.yrfcuj (mgr.14201) 1089 : audit [DBG] from='client.16830 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: audit 2026-03-07T07:14:32.891319+0000 mgr.vm07.yrfcuj (mgr.14201) 1089 : audit [DBG] from='client.16830 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: cluster 2026-03-07T07:14:33.009495+0000 mgr.vm07.yrfcuj (mgr.14201) 1090 : cluster [DBG] pgmap v643: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: cluster 2026-03-07T07:14:33.009495+0000 mgr.vm07.yrfcuj (mgr.14201) 1090 : cluster [DBG] pgmap v643: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: audit 2026-03-07T07:14:33.026665+0000 mon.vm07 (mon.0) 1234 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: audit 2026-03-07T07:14:33.026665+0000 mon.vm07 (mon.0) 1234 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: audit 2026-03-07T07:14:33.031721+0000 mon.vm07 (mon.0) 1235 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: audit 2026-03-07T07:14:33.031721+0000 mon.vm07 (mon.0) 1235 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: audit 2026-03-07T07:14:33.082341+0000 mgr.vm07.yrfcuj (mgr.14201) 1091 : audit [DBG] from='client.16834 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: audit 2026-03-07T07:14:33.082341+0000 mgr.vm07.yrfcuj (mgr.14201) 1091 : audit [DBG] from='client.16834 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: audit 2026-03-07T07:14:33.279801+0000 mon.vm07 (mon.0) 1236 : audit [DBG] from='client.? 192.168.123.107:0/2513058087' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:34.103 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:33 vm07 bash[17031]: audit 2026-03-07T07:14:33.279801+0000 mon.vm07 (mon.0) 1236 : audit [DBG] from='client.? 192.168.123.107:0/2513058087' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: audit 2026-03-07T07:14:32.695137+0000 mgr.vm07.yrfcuj (mgr.14201) 1087 : audit [DBG] from='client.16826 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "rgw.foo.vm09.kpgoql", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: audit 2026-03-07T07:14:32.695137+0000 mgr.vm07.yrfcuj (mgr.14201) 1087 : audit [DBG] from='client.16826 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "rgw.foo.vm09.kpgoql", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: cephadm 2026-03-07T07:14:32.695505+0000 mgr.vm07.yrfcuj (mgr.14201) 1088 : cephadm [INF] Schedule start daemon rgw.foo.vm09.kpgoql 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: cephadm 2026-03-07T07:14:32.695505+0000 mgr.vm07.yrfcuj (mgr.14201) 1088 : cephadm [INF] Schedule start daemon rgw.foo.vm09.kpgoql 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: audit 2026-03-07T07:14:32.700760+0000 mon.vm07 (mon.0) 1231 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: audit 2026-03-07T07:14:32.700760+0000 mon.vm07 (mon.0) 1231 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: audit 2026-03-07T07:14:32.706398+0000 mon.vm07 (mon.0) 1232 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: audit 2026-03-07T07:14:32.706398+0000 mon.vm07 (mon.0) 1232 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: audit 2026-03-07T07:14:32.707397+0000 mon.vm07 (mon.0) 1233 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: audit 2026-03-07T07:14:32.707397+0000 mon.vm07 (mon.0) 1233 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: audit 2026-03-07T07:14:32.891319+0000 mgr.vm07.yrfcuj (mgr.14201) 1089 : audit [DBG] from='client.16830 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: audit 2026-03-07T07:14:32.891319+0000 mgr.vm07.yrfcuj (mgr.14201) 1089 : audit [DBG] from='client.16830 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: cluster 2026-03-07T07:14:33.009495+0000 mgr.vm07.yrfcuj (mgr.14201) 1090 : cluster [DBG] pgmap v643: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: cluster 2026-03-07T07:14:33.009495+0000 mgr.vm07.yrfcuj (mgr.14201) 1090 : cluster [DBG] pgmap v643: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: audit 2026-03-07T07:14:33.026665+0000 mon.vm07 (mon.0) 1234 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: audit 2026-03-07T07:14:33.026665+0000 mon.vm07 (mon.0) 1234 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: audit 2026-03-07T07:14:33.031721+0000 mon.vm07 (mon.0) 1235 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: audit 2026-03-07T07:14:33.031721+0000 mon.vm07 (mon.0) 1235 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: audit 2026-03-07T07:14:33.082341+0000 mgr.vm07.yrfcuj (mgr.14201) 1091 : audit [DBG] from='client.16834 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: audit 2026-03-07T07:14:33.082341+0000 mgr.vm07.yrfcuj (mgr.14201) 1091 : audit [DBG] from='client.16834 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: audit 2026-03-07T07:14:33.279801+0000 mon.vm07 (mon.0) 1236 : audit [DBG] from='client.? 192.168.123.107:0/2513058087' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:34.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:33 vm09 bash[21220]: audit 2026-03-07T07:14:33.279801+0000 mon.vm07 (mon.0) 1236 : audit [DBG] from='client.? 192.168.123.107:0/2513058087' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:36.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:36 vm09 bash[21220]: cluster 2026-03-07T07:14:35.009912+0000 mgr.vm07.yrfcuj (mgr.14201) 1092 : cluster [DBG] pgmap v644: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:14:36.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:36 vm09 bash[21220]: cluster 2026-03-07T07:14:35.009912+0000 mgr.vm07.yrfcuj (mgr.14201) 1092 : cluster [DBG] pgmap v644: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:14:36.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:36 vm07 bash[17031]: cluster 2026-03-07T07:14:35.009912+0000 mgr.vm07.yrfcuj (mgr.14201) 1092 : cluster [DBG] pgmap v644: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:14:36.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:36 vm07 bash[17031]: cluster 2026-03-07T07:14:35.009912+0000 mgr.vm07.yrfcuj (mgr.14201) 1092 : cluster [DBG] pgmap v644: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:14:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:38 vm09 bash[21220]: cluster 2026-03-07T07:14:37.010259+0000 mgr.vm07.yrfcuj (mgr.14201) 1093 : cluster [DBG] pgmap v645: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:14:38.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:38 vm09 bash[21220]: cluster 2026-03-07T07:14:37.010259+0000 mgr.vm07.yrfcuj (mgr.14201) 1093 : cluster [DBG] pgmap v645: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:14:38.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:38 vm07 bash[17031]: cluster 2026-03-07T07:14:37.010259+0000 mgr.vm07.yrfcuj (mgr.14201) 1093 : cluster [DBG] pgmap v645: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:14:38.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:38 vm07 bash[17031]: cluster 2026-03-07T07:14:37.010259+0000 mgr.vm07.yrfcuj (mgr.14201) 1093 : cluster [DBG] pgmap v645: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:14:38.483 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to start 2026-03-07T08:14:38.643 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:14:38.643 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (11m) 0s ago 17m 124M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:14:38.643 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (10m) 0s ago 16m 124M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:14:38.643 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (5m) 4m ago 17m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:14:38.643 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 4m ago 16m - - 2026-03-07T08:14:38.827 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:14:38.827 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:14:38.827 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: audit 2026-03-07T07:14:38.307677+0000 mon.vm07 (mon.0) 1237 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: audit 2026-03-07T07:14:38.307677+0000 mon.vm07 (mon.0) 1237 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: audit 2026-03-07T07:14:38.313736+0000 mon.vm07 (mon.0) 1238 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: audit 2026-03-07T07:14:38.313736+0000 mon.vm07 (mon.0) 1238 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: audit 2026-03-07T07:14:38.314768+0000 mon.vm07 (mon.0) 1239 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: audit 2026-03-07T07:14:38.314768+0000 mon.vm07 (mon.0) 1239 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: audit 2026-03-07T07:14:38.315337+0000 mon.vm07 (mon.0) 1240 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: audit 2026-03-07T07:14:38.315337+0000 mon.vm07 (mon.0) 1240 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: cluster 2026-03-07T07:14:38.316388+0000 mgr.vm07.yrfcuj (mgr.14201) 1094 : cluster [DBG] pgmap v646: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 90 B/s rd, 181 B/s wr, 0 op/s 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: cluster 2026-03-07T07:14:38.316388+0000 mgr.vm07.yrfcuj (mgr.14201) 1094 : cluster [DBG] pgmap v646: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 90 B/s rd, 181 B/s wr, 0 op/s 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: cluster 2026-03-07T07:14:38.316501+0000 mgr.vm07.yrfcuj (mgr.14201) 1095 : cluster [DBG] pgmap v647: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: cluster 2026-03-07T07:14:38.316501+0000 mgr.vm07.yrfcuj (mgr.14201) 1095 : cluster [DBG] pgmap v647: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: audit 2026-03-07T07:14:38.321081+0000 mon.vm07 (mon.0) 1241 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: audit 2026-03-07T07:14:38.321081+0000 mon.vm07 (mon.0) 1241 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: audit 2026-03-07T07:14:38.323012+0000 mon.vm07 (mon.0) 1242 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: audit 2026-03-07T07:14:38.323012+0000 mon.vm07 (mon.0) 1242 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: audit 2026-03-07T07:14:38.471410+0000 mgr.vm07.yrfcuj (mgr.14201) 1096 : audit [DBG] from='client.16842 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: audit 2026-03-07T07:14:38.471410+0000 mgr.vm07.yrfcuj (mgr.14201) 1096 : audit [DBG] from='client.16842 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: audit 2026-03-07T07:14:38.833010+0000 mon.vm07 (mon.0) 1243 : audit [DBG] from='client.? 192.168.123.107:0/3476937492' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:39.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:39 vm09 bash[21220]: audit 2026-03-07T07:14:38.833010+0000 mon.vm07 (mon.0) 1243 : audit [DBG] from='client.? 192.168.123.107:0/3476937492' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: audit 2026-03-07T07:14:38.307677+0000 mon.vm07 (mon.0) 1237 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: audit 2026-03-07T07:14:38.307677+0000 mon.vm07 (mon.0) 1237 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: audit 2026-03-07T07:14:38.313736+0000 mon.vm07 (mon.0) 1238 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: audit 2026-03-07T07:14:38.313736+0000 mon.vm07 (mon.0) 1238 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: audit 2026-03-07T07:14:38.314768+0000 mon.vm07 (mon.0) 1239 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: audit 2026-03-07T07:14:38.314768+0000 mon.vm07 (mon.0) 1239 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: audit 2026-03-07T07:14:38.315337+0000 mon.vm07 (mon.0) 1240 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: audit 2026-03-07T07:14:38.315337+0000 mon.vm07 (mon.0) 1240 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: cluster 2026-03-07T07:14:38.316388+0000 mgr.vm07.yrfcuj (mgr.14201) 1094 : cluster [DBG] pgmap v646: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 90 B/s rd, 181 B/s wr, 0 op/s 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: cluster 2026-03-07T07:14:38.316388+0000 mgr.vm07.yrfcuj (mgr.14201) 1094 : cluster [DBG] pgmap v646: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 90 B/s rd, 181 B/s wr, 0 op/s 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: cluster 2026-03-07T07:14:38.316501+0000 mgr.vm07.yrfcuj (mgr.14201) 1095 : cluster [DBG] pgmap v647: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: cluster 2026-03-07T07:14:38.316501+0000 mgr.vm07.yrfcuj (mgr.14201) 1095 : cluster [DBG] pgmap v647: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: audit 2026-03-07T07:14:38.321081+0000 mon.vm07 (mon.0) 1241 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: audit 2026-03-07T07:14:38.321081+0000 mon.vm07 (mon.0) 1241 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: audit 2026-03-07T07:14:38.323012+0000 mon.vm07 (mon.0) 1242 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: audit 2026-03-07T07:14:38.323012+0000 mon.vm07 (mon.0) 1242 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: audit 2026-03-07T07:14:38.471410+0000 mgr.vm07.yrfcuj (mgr.14201) 1096 : audit [DBG] from='client.16842 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: audit 2026-03-07T07:14:38.471410+0000 mgr.vm07.yrfcuj (mgr.14201) 1096 : audit [DBG] from='client.16842 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: audit 2026-03-07T07:14:38.833010+0000 mon.vm07 (mon.0) 1243 : audit [DBG] from='client.? 192.168.123.107:0/3476937492' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:39.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:39 vm07 bash[17031]: audit 2026-03-07T07:14:38.833010+0000 mon.vm07 (mon.0) 1243 : audit [DBG] from='client.? 192.168.123.107:0/3476937492' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:40.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:40 vm09 bash[21220]: audit 2026-03-07T07:14:38.646408+0000 mgr.vm07.yrfcuj (mgr.14201) 1097 : audit [DBG] from='client.16846 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:40.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:40 vm09 bash[21220]: audit 2026-03-07T07:14:38.646408+0000 mgr.vm07.yrfcuj (mgr.14201) 1097 : audit [DBG] from='client.16846 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:40.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:40 vm09 bash[21220]: audit 2026-03-07T07:14:40.146081+0000 mon.vm07 (mon.0) 1244 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:14:40.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:40 vm09 bash[21220]: audit 2026-03-07T07:14:40.146081+0000 mon.vm07 (mon.0) 1244 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:14:40.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:40 vm07 bash[17031]: audit 2026-03-07T07:14:38.646408+0000 mgr.vm07.yrfcuj (mgr.14201) 1097 : audit [DBG] from='client.16846 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:40.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:40 vm07 bash[17031]: audit 2026-03-07T07:14:38.646408+0000 mgr.vm07.yrfcuj (mgr.14201) 1097 : audit [DBG] from='client.16846 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:40.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:40 vm07 bash[17031]: audit 2026-03-07T07:14:40.146081+0000 mon.vm07 (mon.0) 1244 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:14:40.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:40 vm07 bash[17031]: audit 2026-03-07T07:14:40.146081+0000 mon.vm07 (mon.0) 1244 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:14:41.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:41 vm07 bash[17031]: cluster 2026-03-07T07:14:40.317042+0000 mgr.vm07.yrfcuj (mgr.14201) 1098 : cluster [DBG] pgmap v648: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:41.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:41 vm07 bash[17031]: cluster 2026-03-07T07:14:40.317042+0000 mgr.vm07.yrfcuj (mgr.14201) 1098 : cluster [DBG] pgmap v648: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:41.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:41 vm09 bash[21220]: cluster 2026-03-07T07:14:40.317042+0000 mgr.vm07.yrfcuj (mgr.14201) 1098 : cluster [DBG] pgmap v648: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:41.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:41 vm09 bash[21220]: cluster 2026-03-07T07:14:40.317042+0000 mgr.vm07.yrfcuj (mgr.14201) 1098 : cluster [DBG] pgmap v648: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:43.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:43 vm07 bash[17031]: cluster 2026-03-07T07:14:42.317462+0000 mgr.vm07.yrfcuj (mgr.14201) 1099 : cluster [DBG] pgmap v649: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:43.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:43 vm07 bash[17031]: cluster 2026-03-07T07:14:42.317462+0000 mgr.vm07.yrfcuj (mgr.14201) 1099 : cluster [DBG] pgmap v649: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:43.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:43 vm09 bash[21220]: cluster 2026-03-07T07:14:42.317462+0000 mgr.vm07.yrfcuj (mgr.14201) 1099 : cluster [DBG] pgmap v649: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:43.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:43 vm09 bash[21220]: cluster 2026-03-07T07:14:42.317462+0000 mgr.vm07.yrfcuj (mgr.14201) 1099 : cluster [DBG] pgmap v649: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:44.005 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to start 2026-03-07T08:14:44.164 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:14:44.164 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (11m) 5s ago 17m 124M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:14:44.164 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (10m) 5s ago 17m 124M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:14:44.164 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (5m) 4m ago 17m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:14:44.164 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 4m ago 17m - - 2026-03-07T08:14:44.376 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:14:44.376 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:14:44.376 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:14:44.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:44 vm09 bash[21220]: audit 2026-03-07T07:14:44.382058+0000 mon.vm07 (mon.0) 1245 : audit [DBG] from='client.? 192.168.123.107:0/2882800499' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:44.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:44 vm09 bash[21220]: audit 2026-03-07T07:14:44.382058+0000 mon.vm07 (mon.0) 1245 : audit [DBG] from='client.? 192.168.123.107:0/2882800499' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:44.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:44 vm07 bash[17031]: audit 2026-03-07T07:14:44.382058+0000 mon.vm07 (mon.0) 1245 : audit [DBG] from='client.? 192.168.123.107:0/2882800499' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:44.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:44 vm07 bash[17031]: audit 2026-03-07T07:14:44.382058+0000 mon.vm07 (mon.0) 1245 : audit [DBG] from='client.? 192.168.123.107:0/2882800499' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:45 vm09 bash[21220]: audit 2026-03-07T07:14:43.995612+0000 mgr.vm07.yrfcuj (mgr.14201) 1100 : audit [DBG] from='client.16854 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:45 vm09 bash[21220]: audit 2026-03-07T07:14:43.995612+0000 mgr.vm07.yrfcuj (mgr.14201) 1100 : audit [DBG] from='client.16854 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:45 vm09 bash[21220]: audit 2026-03-07T07:14:44.167238+0000 mgr.vm07.yrfcuj (mgr.14201) 1101 : audit [DBG] from='client.16858 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:45 vm09 bash[21220]: audit 2026-03-07T07:14:44.167238+0000 mgr.vm07.yrfcuj (mgr.14201) 1101 : audit [DBG] from='client.16858 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:45 vm09 bash[21220]: cluster 2026-03-07T07:14:44.317900+0000 mgr.vm07.yrfcuj (mgr.14201) 1102 : cluster [DBG] pgmap v650: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:45.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:45 vm09 bash[21220]: cluster 2026-03-07T07:14:44.317900+0000 mgr.vm07.yrfcuj (mgr.14201) 1102 : cluster [DBG] pgmap v650: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:45.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:45 vm07 bash[17031]: audit 2026-03-07T07:14:43.995612+0000 mgr.vm07.yrfcuj (mgr.14201) 1100 : audit [DBG] from='client.16854 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:45.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:45 vm07 bash[17031]: audit 2026-03-07T07:14:43.995612+0000 mgr.vm07.yrfcuj (mgr.14201) 1100 : audit [DBG] from='client.16854 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:45.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:45 vm07 bash[17031]: audit 2026-03-07T07:14:44.167238+0000 mgr.vm07.yrfcuj (mgr.14201) 1101 : audit [DBG] from='client.16858 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:45.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:45 vm07 bash[17031]: audit 2026-03-07T07:14:44.167238+0000 mgr.vm07.yrfcuj (mgr.14201) 1101 : audit [DBG] from='client.16858 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:45.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:45 vm07 bash[17031]: cluster 2026-03-07T07:14:44.317900+0000 mgr.vm07.yrfcuj (mgr.14201) 1102 : cluster [DBG] pgmap v650: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:45.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:45 vm07 bash[17031]: cluster 2026-03-07T07:14:44.317900+0000 mgr.vm07.yrfcuj (mgr.14201) 1102 : cluster [DBG] pgmap v650: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:47 vm09 bash[21220]: cluster 2026-03-07T07:14:46.318265+0000 mgr.vm07.yrfcuj (mgr.14201) 1103 : cluster [DBG] pgmap v651: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:47 vm09 bash[21220]: cluster 2026-03-07T07:14:46.318265+0000 mgr.vm07.yrfcuj (mgr.14201) 1103 : cluster [DBG] pgmap v651: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:47 vm09 bash[21220]: audit 2026-03-07T07:14:47.206100+0000 mon.vm07 (mon.0) 1246 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:47 vm09 bash[21220]: audit 2026-03-07T07:14:47.206100+0000 mon.vm07 (mon.0) 1246 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:47 vm09 bash[21220]: audit 2026-03-07T07:14:47.212335+0000 mon.vm07 (mon.0) 1247 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:47 vm09 bash[21220]: audit 2026-03-07T07:14:47.212335+0000 mon.vm07 (mon.0) 1247 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:47 vm09 bash[21220]: audit 2026-03-07T07:14:47.241875+0000 mon.vm07 (mon.0) 1248 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:14:47.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:47 vm09 bash[21220]: audit 2026-03-07T07:14:47.241875+0000 mon.vm07 (mon.0) 1248 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:14:47.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:47 vm07 bash[17031]: cluster 2026-03-07T07:14:46.318265+0000 mgr.vm07.yrfcuj (mgr.14201) 1103 : cluster [DBG] pgmap v651: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:47.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:47 vm07 bash[17031]: cluster 2026-03-07T07:14:46.318265+0000 mgr.vm07.yrfcuj (mgr.14201) 1103 : cluster [DBG] pgmap v651: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 110 B/s rd, 220 B/s wr, 0 op/s 2026-03-07T08:14:47.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:47 vm07 bash[17031]: audit 2026-03-07T07:14:47.206100+0000 mon.vm07 (mon.0) 1246 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:47.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:47 vm07 bash[17031]: audit 2026-03-07T07:14:47.206100+0000 mon.vm07 (mon.0) 1246 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:47.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:47 vm07 bash[17031]: audit 2026-03-07T07:14:47.212335+0000 mon.vm07 (mon.0) 1247 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:47.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:47 vm07 bash[17031]: audit 2026-03-07T07:14:47.212335+0000 mon.vm07 (mon.0) 1247 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:47.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:47 vm07 bash[17031]: audit 2026-03-07T07:14:47.241875+0000 mon.vm07 (mon.0) 1248 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:14:47.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:47 vm07 bash[17031]: audit 2026-03-07T07:14:47.241875+0000 mon.vm07 (mon.0) 1248 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:14:49.561 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for rgw.foo.vm09.kpgoql to start 2026-03-07T08:14:49.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:49 vm07 bash[17031]: cluster 2026-03-07T07:14:48.318742+0000 mgr.vm07.yrfcuj (mgr.14201) 1104 : cluster [DBG] pgmap v652: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 7.7 KiB/s rd, 204 B/s wr, 11 op/s 2026-03-07T08:14:49.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:49 vm07 bash[17031]: cluster 2026-03-07T07:14:48.318742+0000 mgr.vm07.yrfcuj (mgr.14201) 1104 : cluster [DBG] pgmap v652: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 7.7 KiB/s rd, 204 B/s wr, 11 op/s 2026-03-07T08:14:49.714 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:14:49.714 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (11m) 11s ago 17m 124M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:14:49.714 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (10m) 11s ago 17m 124M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:14:49.714 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (5m) 4m ago 17m 90.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:14:49.714 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 error 4m ago 17m - - 2026-03-07T08:14:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:49 vm09 bash[21220]: cluster 2026-03-07T07:14:48.318742+0000 mgr.vm07.yrfcuj (mgr.14201) 1104 : cluster [DBG] pgmap v652: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 7.7 KiB/s rd, 204 B/s wr, 11 op/s 2026-03-07T08:14:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:49 vm09 bash[21220]: cluster 2026-03-07T07:14:48.318742+0000 mgr.vm07.yrfcuj (mgr.14201) 1104 : cluster [DBG] pgmap v652: 129 pgs: 129 active+clean; 454 KiB data, 250 MiB used, 160 GiB / 160 GiB avail; 7.7 KiB/s rd, 204 B/s wr, 11 op/s 2026-03-07T08:14:49.908 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_WARN 1 failed cephadm daemon(s) 2026-03-07T08:14:49.908 INFO:teuthology.orchestra.run.vm07.stdout:[WRN] CEPHADM_FAILED_DAEMON: 1 failed cephadm daemon(s) 2026-03-07T08:14:49.908 INFO:teuthology.orchestra.run.vm07.stdout: daemon rgw.foo.vm09.kpgoql on vm09 is in error state 2026-03-07T08:14:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:50 vm09 bash[21220]: audit 2026-03-07T07:14:49.552036+0000 mgr.vm07.yrfcuj (mgr.14201) 1105 : audit [DBG] from='client.16870 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:50 vm09 bash[21220]: audit 2026-03-07T07:14:49.552036+0000 mgr.vm07.yrfcuj (mgr.14201) 1105 : audit [DBG] from='client.16870 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:50 vm09 bash[21220]: audit 2026-03-07T07:14:49.914017+0000 mon.vm07 (mon.0) 1249 : audit [DBG] from='client.? 192.168.123.107:0/837026153' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:50.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:50 vm09 bash[21220]: audit 2026-03-07T07:14:49.914017+0000 mon.vm07 (mon.0) 1249 : audit [DBG] from='client.? 192.168.123.107:0/837026153' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:50.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:50 vm07 bash[17031]: audit 2026-03-07T07:14:49.552036+0000 mgr.vm07.yrfcuj (mgr.14201) 1105 : audit [DBG] from='client.16870 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:50.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:50 vm07 bash[17031]: audit 2026-03-07T07:14:49.552036+0000 mgr.vm07.yrfcuj (mgr.14201) 1105 : audit [DBG] from='client.16870 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:50.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:50 vm07 bash[17031]: audit 2026-03-07T07:14:49.914017+0000 mon.vm07 (mon.0) 1249 : audit [DBG] from='client.? 192.168.123.107:0/837026153' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:50.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:50 vm07 bash[17031]: audit 2026-03-07T07:14:49.914017+0000 mon.vm07 (mon.0) 1249 : audit [DBG] from='client.? 192.168.123.107:0/837026153' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:51.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:51 vm09 bash[21220]: audit 2026-03-07T07:14:49.715914+0000 mgr.vm07.yrfcuj (mgr.14201) 1106 : audit [DBG] from='client.25871 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:51.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:51 vm09 bash[21220]: audit 2026-03-07T07:14:49.715914+0000 mgr.vm07.yrfcuj (mgr.14201) 1106 : audit [DBG] from='client.25871 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:51.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:51 vm09 bash[21220]: cluster 2026-03-07T07:14:50.319274+0000 mgr.vm07.yrfcuj (mgr.14201) 1107 : cluster [DBG] pgmap v653: 129 pgs: 129 active+clean; 454 KiB data, 255 MiB used, 160 GiB / 160 GiB avail; 16 KiB/s rd, 170 B/s wr, 25 op/s 2026-03-07T08:14:51.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:51 vm09 bash[21220]: cluster 2026-03-07T07:14:50.319274+0000 mgr.vm07.yrfcuj (mgr.14201) 1107 : cluster [DBG] pgmap v653: 129 pgs: 129 active+clean; 454 KiB data, 255 MiB used, 160 GiB / 160 GiB avail; 16 KiB/s rd, 170 B/s wr, 25 op/s 2026-03-07T08:14:51.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:51 vm07 bash[17031]: audit 2026-03-07T07:14:49.715914+0000 mgr.vm07.yrfcuj (mgr.14201) 1106 : audit [DBG] from='client.25871 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:51.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:51 vm07 bash[17031]: audit 2026-03-07T07:14:49.715914+0000 mgr.vm07.yrfcuj (mgr.14201) 1106 : audit [DBG] from='client.25871 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "rgw", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:51.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:51 vm07 bash[17031]: cluster 2026-03-07T07:14:50.319274+0000 mgr.vm07.yrfcuj (mgr.14201) 1107 : cluster [DBG] pgmap v653: 129 pgs: 129 active+clean; 454 KiB data, 255 MiB used, 160 GiB / 160 GiB avail; 16 KiB/s rd, 170 B/s wr, 25 op/s 2026-03-07T08:14:51.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:51 vm07 bash[17031]: cluster 2026-03-07T07:14:50.319274+0000 mgr.vm07.yrfcuj (mgr.14201) 1107 : cluster [DBG] pgmap v653: 129 pgs: 129 active+clean; 454 KiB data, 255 MiB used, 160 GiB / 160 GiB avail; 16 KiB/s rd, 170 B/s wr, 25 op/s 2026-03-07T08:14:52.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:52 vm07 bash[17031]: audit 2026-03-07T07:14:51.647291+0000 mon.vm07 (mon.0) 1250 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:52.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:52 vm07 bash[17031]: audit 2026-03-07T07:14:51.647291+0000 mon.vm07 (mon.0) 1250 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:52.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:52 vm07 bash[17031]: audit 2026-03-07T07:14:51.653159+0000 mon.vm07 (mon.0) 1251 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:52.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:52 vm07 bash[17031]: audit 2026-03-07T07:14:51.653159+0000 mon.vm07 (mon.0) 1251 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:52.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:52 vm07 bash[17031]: audit 2026-03-07T07:14:51.654151+0000 mon.vm07 (mon.0) 1252 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:14:52.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:52 vm07 bash[17031]: audit 2026-03-07T07:14:51.654151+0000 mon.vm07 (mon.0) 1252 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:14:52.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:52 vm07 bash[17031]: audit 2026-03-07T07:14:51.654621+0000 mon.vm07 (mon.0) 1253 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:14:52.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:52 vm07 bash[17031]: audit 2026-03-07T07:14:51.654621+0000 mon.vm07 (mon.0) 1253 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:14:52.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:52 vm07 bash[17031]: cluster 2026-03-07T07:14:51.655595+0000 mgr.vm07.yrfcuj (mgr.14201) 1108 : cluster [DBG] pgmap v654: 129 pgs: 129 active+clean; 454 KiB data, 255 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s rd, 180 B/s wr, 41 op/s 2026-03-07T08:14:52.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:52 vm07 bash[17031]: cluster 2026-03-07T07:14:51.655595+0000 mgr.vm07.yrfcuj (mgr.14201) 1108 : cluster [DBG] pgmap v654: 129 pgs: 129 active+clean; 454 KiB data, 255 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s rd, 180 B/s wr, 41 op/s 2026-03-07T08:14:52.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:52 vm07 bash[17031]: audit 2026-03-07T07:14:51.659479+0000 mon.vm07 (mon.0) 1254 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:52.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:52 vm07 bash[17031]: audit 2026-03-07T07:14:51.659479+0000 mon.vm07 (mon.0) 1254 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:52.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:52 vm07 bash[17031]: audit 2026-03-07T07:14:51.661027+0000 mon.vm07 (mon.0) 1255 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:14:52.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:52 vm07 bash[17031]: audit 2026-03-07T07:14:51.661027+0000 mon.vm07 (mon.0) 1255 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:14:53.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:52 vm09 bash[21220]: audit 2026-03-07T07:14:51.647291+0000 mon.vm07 (mon.0) 1250 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:53.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:52 vm09 bash[21220]: audit 2026-03-07T07:14:51.647291+0000 mon.vm07 (mon.0) 1250 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:53.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:52 vm09 bash[21220]: audit 2026-03-07T07:14:51.653159+0000 mon.vm07 (mon.0) 1251 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:53.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:52 vm09 bash[21220]: audit 2026-03-07T07:14:51.653159+0000 mon.vm07 (mon.0) 1251 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:53.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:52 vm09 bash[21220]: audit 2026-03-07T07:14:51.654151+0000 mon.vm07 (mon.0) 1252 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:14:53.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:52 vm09 bash[21220]: audit 2026-03-07T07:14:51.654151+0000 mon.vm07 (mon.0) 1252 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:14:53.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:52 vm09 bash[21220]: audit 2026-03-07T07:14:51.654621+0000 mon.vm07 (mon.0) 1253 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:14:53.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:52 vm09 bash[21220]: audit 2026-03-07T07:14:51.654621+0000 mon.vm07 (mon.0) 1253 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:14:53.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:52 vm09 bash[21220]: cluster 2026-03-07T07:14:51.655595+0000 mgr.vm07.yrfcuj (mgr.14201) 1108 : cluster [DBG] pgmap v654: 129 pgs: 129 active+clean; 454 KiB data, 255 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s rd, 180 B/s wr, 41 op/s 2026-03-07T08:14:53.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:52 vm09 bash[21220]: cluster 2026-03-07T07:14:51.655595+0000 mgr.vm07.yrfcuj (mgr.14201) 1108 : cluster [DBG] pgmap v654: 129 pgs: 129 active+clean; 454 KiB data, 255 MiB used, 160 GiB / 160 GiB avail; 27 KiB/s rd, 180 B/s wr, 41 op/s 2026-03-07T08:14:53.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:52 vm09 bash[21220]: audit 2026-03-07T07:14:51.659479+0000 mon.vm07 (mon.0) 1254 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:53.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:52 vm09 bash[21220]: audit 2026-03-07T07:14:51.659479+0000 mon.vm07 (mon.0) 1254 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:53.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:52 vm09 bash[21220]: audit 2026-03-07T07:14:51.661027+0000 mon.vm07 (mon.0) 1255 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:14:53.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:52 vm09 bash[21220]: audit 2026-03-07T07:14:51.661027+0000 mon.vm07 (mon.0) 1255 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:14:53.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:53 vm07 bash[17031]: cluster 2026-03-07T07:14:52.663006+0000 mon.vm07 (mon.0) 1256 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-07T08:14:53.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:53 vm07 bash[17031]: cluster 2026-03-07T07:14:52.663006+0000 mon.vm07 (mon.0) 1256 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-07T08:14:53.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:53 vm07 bash[17031]: cluster 2026-03-07T07:14:52.663014+0000 mon.vm07 (mon.0) 1257 : cluster [INF] Cluster is now healthy 2026-03-07T08:14:53.898 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:53 vm07 bash[17031]: cluster 2026-03-07T07:14:52.663014+0000 mon.vm07 (mon.0) 1257 : cluster [INF] Cluster is now healthy 2026-03-07T08:14:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:53 vm09 bash[21220]: cluster 2026-03-07T07:14:52.663006+0000 mon.vm07 (mon.0) 1256 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-07T08:14:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:53 vm09 bash[21220]: cluster 2026-03-07T07:14:52.663006+0000 mon.vm07 (mon.0) 1256 : cluster [INF] Health check cleared: CEPHADM_FAILED_DAEMON (was: 1 failed cephadm daemon(s)) 2026-03-07T08:14:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:53 vm09 bash[21220]: cluster 2026-03-07T07:14:52.663014+0000 mon.vm07 (mon.0) 1257 : cluster [INF] Cluster is now healthy 2026-03-07T08:14:54.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:53 vm09 bash[21220]: cluster 2026-03-07T07:14:52.663014+0000 mon.vm07 (mon.0) 1257 : cluster [INF] Cluster is now healthy 2026-03-07T08:14:55.086 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (7s) 3s ago 17m 89.2M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 8ac366e53c02 2026-03-07T08:14:55.086 INFO:teuthology.orchestra.run.vm07.stdout:Check with each haproxy down in turn... 2026-03-07T08:14:55.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:54 vm09 bash[21220]: cluster 2026-03-07T07:14:53.655971+0000 mgr.vm07.yrfcuj (mgr.14201) 1109 : cluster [DBG] pgmap v655: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 0 B/s wr, 86 op/s 2026-03-07T08:14:55.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:54 vm09 bash[21220]: cluster 2026-03-07T07:14:53.655971+0000 mgr.vm07.yrfcuj (mgr.14201) 1109 : cluster [DBG] pgmap v655: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 0 B/s wr, 86 op/s 2026-03-07T08:14:55.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:54 vm07 bash[17031]: cluster 2026-03-07T07:14:53.655971+0000 mgr.vm07.yrfcuj (mgr.14201) 1109 : cluster [DBG] pgmap v655: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 0 B/s wr, 86 op/s 2026-03-07T08:14:55.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:54 vm07 bash[17031]: cluster 2026-03-07T07:14:53.655971+0000 mgr.vm07.yrfcuj (mgr.14201) 1109 : cluster [DBG] pgmap v655: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 0 B/s wr, 86 op/s 2026-03-07T08:14:55.446 INFO:teuthology.orchestra.run.vm07.stdout:Scheduled to stop haproxy.rgw.foo.vm07.rzyepz on host 'vm07' 2026-03-07T08:14:55.648 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for haproxy.rgw.foo.vm07.rzyepz to stop 2026-03-07T08:14:55.859 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:14:55.859 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm07.rzyepz vm07 *:9000,9001 running (17m) 17s ago 17m 3691k - 2.3.17-d1c9119 e85424b0d443 67deea3aeb4a 2026-03-07T08:14:55.859 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm09.hhbnqk vm09 *:9000,9001 running (17m) 4s ago 17m 3663k - 2.3.17-d1c9119 e85424b0d443 7325e33789e3 2026-03-07T08:14:56.069 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.076963+0000 mgr.vm07.yrfcuj (mgr.14201) 1110 : audit [DBG] from='client.16882 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.076963+0000 mgr.vm07.yrfcuj (mgr.14201) 1110 : audit [DBG] from='client.16882 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.152052+0000 mon.vm07 (mon.0) 1258 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.152052+0000 mon.vm07 (mon.0) 1258 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.153287+0000 mon.vm07 (mon.0) 1259 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.153287+0000 mon.vm07 (mon.0) 1259 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.268899+0000 mgr.vm07.yrfcuj (mgr.14201) 1111 : audit [DBG] from='client.16886 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.268899+0000 mgr.vm07.yrfcuj (mgr.14201) 1111 : audit [DBG] from='client.16886 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.439602+0000 mgr.vm07.yrfcuj (mgr.14201) 1112 : audit [DBG] from='client.25885 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "haproxy.rgw.foo.vm07.rzyepz", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.439602+0000 mgr.vm07.yrfcuj (mgr.14201) 1112 : audit [DBG] from='client.25885 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "haproxy.rgw.foo.vm07.rzyepz", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: cephadm 2026-03-07T07:14:55.440065+0000 mgr.vm07.yrfcuj (mgr.14201) 1113 : cephadm [INF] Schedule stop daemon haproxy.rgw.foo.vm07.rzyepz 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: cephadm 2026-03-07T07:14:55.440065+0000 mgr.vm07.yrfcuj (mgr.14201) 1113 : cephadm [INF] Schedule stop daemon haproxy.rgw.foo.vm07.rzyepz 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.446165+0000 mon.vm07 (mon.0) 1260 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.446165+0000 mon.vm07 (mon.0) 1260 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.451324+0000 mon.vm07 (mon.0) 1261 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.451324+0000 mon.vm07 (mon.0) 1261 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.452608+0000 mon.vm07 (mon.0) 1262 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.452608+0000 mon.vm07 (mon.0) 1262 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.636728+0000 mgr.vm07.yrfcuj (mgr.14201) 1114 : audit [DBG] from='client.16894 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.636728+0000 mgr.vm07.yrfcuj (mgr.14201) 1114 : audit [DBG] from='client.16894 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.817900+0000 mon.vm07 (mon.0) 1263 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.817900+0000 mon.vm07 (mon.0) 1263 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.818411+0000 mon.vm07 (mon.0) 1264 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.818411+0000 mon.vm07 (mon.0) 1264 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.823674+0000 mon.vm07 (mon.0) 1265 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.823674+0000 mon.vm07 (mon.0) 1265 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:56.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.825638+0000 mon.vm07 (mon.0) 1266 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:14:56.616 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:55.825638+0000 mon.vm07 (mon.0) 1266 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:14:56.616 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:56.074753+0000 mon.vm07 (mon.0) 1267 : audit [DBG] from='client.? 192.168.123.107:0/1319891440' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:56.616 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:56 vm09 bash[21220]: audit 2026-03-07T07:14:56.074753+0000 mon.vm07 (mon.0) 1267 : audit [DBG] from='client.? 192.168.123.107:0/1319891440' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.076963+0000 mgr.vm07.yrfcuj (mgr.14201) 1110 : audit [DBG] from='client.16882 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.076963+0000 mgr.vm07.yrfcuj (mgr.14201) 1110 : audit [DBG] from='client.16882 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.152052+0000 mon.vm07 (mon.0) 1258 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.152052+0000 mon.vm07 (mon.0) 1258 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.153287+0000 mon.vm07 (mon.0) 1259 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.153287+0000 mon.vm07 (mon.0) 1259 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.268899+0000 mgr.vm07.yrfcuj (mgr.14201) 1111 : audit [DBG] from='client.16886 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.268899+0000 mgr.vm07.yrfcuj (mgr.14201) 1111 : audit [DBG] from='client.16886 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.439602+0000 mgr.vm07.yrfcuj (mgr.14201) 1112 : audit [DBG] from='client.25885 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "haproxy.rgw.foo.vm07.rzyepz", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.439602+0000 mgr.vm07.yrfcuj (mgr.14201) 1112 : audit [DBG] from='client.25885 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "haproxy.rgw.foo.vm07.rzyepz", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: cephadm 2026-03-07T07:14:55.440065+0000 mgr.vm07.yrfcuj (mgr.14201) 1113 : cephadm [INF] Schedule stop daemon haproxy.rgw.foo.vm07.rzyepz 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: cephadm 2026-03-07T07:14:55.440065+0000 mgr.vm07.yrfcuj (mgr.14201) 1113 : cephadm [INF] Schedule stop daemon haproxy.rgw.foo.vm07.rzyepz 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.446165+0000 mon.vm07 (mon.0) 1260 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.446165+0000 mon.vm07 (mon.0) 1260 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.451324+0000 mon.vm07 (mon.0) 1261 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.451324+0000 mon.vm07 (mon.0) 1261 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.452608+0000 mon.vm07 (mon.0) 1262 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.452608+0000 mon.vm07 (mon.0) 1262 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.636728+0000 mgr.vm07.yrfcuj (mgr.14201) 1114 : audit [DBG] from='client.16894 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.636728+0000 mgr.vm07.yrfcuj (mgr.14201) 1114 : audit [DBG] from='client.16894 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.817900+0000 mon.vm07 (mon.0) 1263 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.817900+0000 mon.vm07 (mon.0) 1263 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.818411+0000 mon.vm07 (mon.0) 1264 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.818411+0000 mon.vm07 (mon.0) 1264 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.823674+0000 mon.vm07 (mon.0) 1265 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.823674+0000 mon.vm07 (mon.0) 1265 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.825638+0000 mon.vm07 (mon.0) 1266 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:55.825638+0000 mon.vm07 (mon.0) 1266 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:56.074753+0000 mon.vm07 (mon.0) 1267 : audit [DBG] from='client.? 192.168.123.107:0/1319891440' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:56.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:56 vm07 bash[17031]: audit 2026-03-07T07:14:56.074753+0000 mon.vm07 (mon.0) 1267 : audit [DBG] from='client.? 192.168.123.107:0/1319891440' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:14:57.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:57 vm09 bash[21220]: cluster 2026-03-07T07:14:55.656402+0000 mgr.vm07.yrfcuj (mgr.14201) 1115 : cluster [DBG] pgmap v656: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 0 B/s wr, 86 op/s 2026-03-07T08:14:57.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:57 vm09 bash[21220]: cluster 2026-03-07T07:14:55.656402+0000 mgr.vm07.yrfcuj (mgr.14201) 1115 : cluster [DBG] pgmap v656: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 0 B/s wr, 86 op/s 2026-03-07T08:14:57.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:57 vm09 bash[21220]: audit 2026-03-07T07:14:55.851403+0000 mgr.vm07.yrfcuj (mgr.14201) 1116 : audit [DBG] from='client.16898 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:57.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:57 vm09 bash[21220]: audit 2026-03-07T07:14:55.851403+0000 mgr.vm07.yrfcuj (mgr.14201) 1116 : audit [DBG] from='client.16898 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:57.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:57 vm07 bash[17031]: cluster 2026-03-07T07:14:55.656402+0000 mgr.vm07.yrfcuj (mgr.14201) 1115 : cluster [DBG] pgmap v656: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 0 B/s wr, 86 op/s 2026-03-07T08:14:57.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:57 vm07 bash[17031]: cluster 2026-03-07T07:14:55.656402+0000 mgr.vm07.yrfcuj (mgr.14201) 1115 : cluster [DBG] pgmap v656: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 0 B/s wr, 86 op/s 2026-03-07T08:14:57.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:57 vm07 bash[17031]: audit 2026-03-07T07:14:55.851403+0000 mgr.vm07.yrfcuj (mgr.14201) 1116 : audit [DBG] from='client.16898 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:57.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:57 vm07 bash[17031]: audit 2026-03-07T07:14:55.851403+0000 mgr.vm07.yrfcuj (mgr.14201) 1116 : audit [DBG] from='client.16898 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:14:59.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:59 vm09 bash[21220]: cluster 2026-03-07T07:14:57.656796+0000 mgr.vm07.yrfcuj (mgr.14201) 1117 : cluster [DBG] pgmap v657: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 180 B/s wr, 87 op/s 2026-03-07T08:14:59.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:14:59 vm09 bash[21220]: cluster 2026-03-07T07:14:57.656796+0000 mgr.vm07.yrfcuj (mgr.14201) 1117 : cluster [DBG] pgmap v657: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 180 B/s wr, 87 op/s 2026-03-07T08:14:59.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:59 vm07 bash[17031]: cluster 2026-03-07T07:14:57.656796+0000 mgr.vm07.yrfcuj (mgr.14201) 1117 : cluster [DBG] pgmap v657: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 180 B/s wr, 87 op/s 2026-03-07T08:14:59.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:14:59 vm07 bash[17031]: cluster 2026-03-07T07:14:57.656796+0000 mgr.vm07.yrfcuj (mgr.14201) 1117 : cluster [DBG] pgmap v657: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 56 KiB/s rd, 180 B/s wr, 87 op/s 2026-03-07T08:15:01.283 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for haproxy.rgw.foo.vm07.rzyepz to stop 2026-03-07T08:15:01.446 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:15:01.446 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm07.rzyepz vm07 *:9000,9001 running (17m) 23s ago 17m 3691k - 2.3.17-d1c9119 e85424b0d443 67deea3aeb4a 2026-03-07T08:15:01.446 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm09.hhbnqk vm09 *:9000,9001 running (17m) 9s ago 17m 3663k - 2.3.17-d1c9119 e85424b0d443 7325e33789e3 2026-03-07T08:15:01.674 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:15:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:01 vm09 bash[21220]: cluster 2026-03-07T07:14:59.657123+0000 mgr.vm07.yrfcuj (mgr.14201) 1118 : cluster [DBG] pgmap v658: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 49 KiB/s rd, 180 B/s wr, 76 op/s 2026-03-07T08:15:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:01 vm09 bash[21220]: cluster 2026-03-07T07:14:59.657123+0000 mgr.vm07.yrfcuj (mgr.14201) 1118 : cluster [DBG] pgmap v658: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 49 KiB/s rd, 180 B/s wr, 76 op/s 2026-03-07T08:15:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:01 vm09 bash[21220]: audit 2026-03-07T07:15:00.812219+0000 mon.vm07 (mon.0) 1268 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:01 vm09 bash[21220]: audit 2026-03-07T07:15:00.812219+0000 mon.vm07 (mon.0) 1268 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:01 vm09 bash[21220]: audit 2026-03-07T07:15:00.817062+0000 mon.vm07 (mon.0) 1269 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:01 vm09 bash[21220]: audit 2026-03-07T07:15:00.817062+0000 mon.vm07 (mon.0) 1269 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:01 vm09 bash[21220]: audit 2026-03-07T07:15:00.846841+0000 mon.vm07 (mon.0) 1270 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:01 vm09 bash[21220]: audit 2026-03-07T07:15:00.846841+0000 mon.vm07 (mon.0) 1270 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:01.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:01 vm07 bash[17031]: cluster 2026-03-07T07:14:59.657123+0000 mgr.vm07.yrfcuj (mgr.14201) 1118 : cluster [DBG] pgmap v658: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 49 KiB/s rd, 180 B/s wr, 76 op/s 2026-03-07T08:15:01.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:01 vm07 bash[17031]: cluster 2026-03-07T07:14:59.657123+0000 mgr.vm07.yrfcuj (mgr.14201) 1118 : cluster [DBG] pgmap v658: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 49 KiB/s rd, 180 B/s wr, 76 op/s 2026-03-07T08:15:01.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:01 vm07 bash[17031]: audit 2026-03-07T07:15:00.812219+0000 mon.vm07 (mon.0) 1268 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:01.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:01 vm07 bash[17031]: audit 2026-03-07T07:15:00.812219+0000 mon.vm07 (mon.0) 1268 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:01.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:01 vm07 bash[17031]: audit 2026-03-07T07:15:00.817062+0000 mon.vm07 (mon.0) 1269 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:01.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:01 vm07 bash[17031]: audit 2026-03-07T07:15:00.817062+0000 mon.vm07 (mon.0) 1269 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:01.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:01 vm07 bash[17031]: audit 2026-03-07T07:15:00.846841+0000 mon.vm07 (mon.0) 1270 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:01.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:01 vm07 bash[17031]: audit 2026-03-07T07:15:00.846841+0000 mon.vm07 (mon.0) 1270 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:02.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:02 vm09 bash[21220]: audit 2026-03-07T07:15:01.272424+0000 mgr.vm07.yrfcuj (mgr.14201) 1119 : audit [DBG] from='client.16906 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:02.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:02 vm09 bash[21220]: audit 2026-03-07T07:15:01.272424+0000 mgr.vm07.yrfcuj (mgr.14201) 1119 : audit [DBG] from='client.16906 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:02.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:02 vm09 bash[21220]: audit 2026-03-07T07:15:01.449833+0000 mgr.vm07.yrfcuj (mgr.14201) 1120 : audit [DBG] from='client.16910 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:02.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:02 vm09 bash[21220]: audit 2026-03-07T07:15:01.449833+0000 mgr.vm07.yrfcuj (mgr.14201) 1120 : audit [DBG] from='client.16910 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:02.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:02 vm09 bash[21220]: audit 2026-03-07T07:15:01.680328+0000 mon.vm07 (mon.0) 1271 : audit [DBG] from='client.? 192.168.123.107:0/1470999700' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:02.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:02 vm09 bash[21220]: audit 2026-03-07T07:15:01.680328+0000 mon.vm07 (mon.0) 1271 : audit [DBG] from='client.? 192.168.123.107:0/1470999700' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:02.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:02 vm07 bash[17031]: audit 2026-03-07T07:15:01.272424+0000 mgr.vm07.yrfcuj (mgr.14201) 1119 : audit [DBG] from='client.16906 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:02.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:02 vm07 bash[17031]: audit 2026-03-07T07:15:01.272424+0000 mgr.vm07.yrfcuj (mgr.14201) 1119 : audit [DBG] from='client.16906 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:02.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:02 vm07 bash[17031]: audit 2026-03-07T07:15:01.449833+0000 mgr.vm07.yrfcuj (mgr.14201) 1120 : audit [DBG] from='client.16910 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:02.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:02 vm07 bash[17031]: audit 2026-03-07T07:15:01.449833+0000 mgr.vm07.yrfcuj (mgr.14201) 1120 : audit [DBG] from='client.16910 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:02.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:02 vm07 bash[17031]: audit 2026-03-07T07:15:01.680328+0000 mon.vm07 (mon.0) 1271 : audit [DBG] from='client.? 192.168.123.107:0/1470999700' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:02.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:02 vm07 bash[17031]: audit 2026-03-07T07:15:01.680328+0000 mon.vm07 (mon.0) 1271 : audit [DBG] from='client.? 192.168.123.107:0/1470999700' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:03.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:03 vm07 bash[17031]: cluster 2026-03-07T07:15:01.657554+0000 mgr.vm07.yrfcuj (mgr.14201) 1121 : cluster [DBG] pgmap v659: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 39 KiB/s rd, 180 B/s wr, 60 op/s 2026-03-07T08:15:03.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:03 vm07 bash[17031]: cluster 2026-03-07T07:15:01.657554+0000 mgr.vm07.yrfcuj (mgr.14201) 1121 : cluster [DBG] pgmap v659: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 39 KiB/s rd, 180 B/s wr, 60 op/s 2026-03-07T08:15:03.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:03 vm09 bash[21220]: cluster 2026-03-07T07:15:01.657554+0000 mgr.vm07.yrfcuj (mgr.14201) 1121 : cluster [DBG] pgmap v659: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 39 KiB/s rd, 180 B/s wr, 60 op/s 2026-03-07T08:15:03.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:03 vm09 bash[21220]: cluster 2026-03-07T07:15:01.657554+0000 mgr.vm07.yrfcuj (mgr.14201) 1121 : cluster [DBG] pgmap v659: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 39 KiB/s rd, 180 B/s wr, 60 op/s 2026-03-07T08:15:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:05 vm09 bash[21220]: cluster 2026-03-07T07:15:03.657956+0000 mgr.vm07.yrfcuj (mgr.14201) 1122 : cluster [DBG] pgmap v660: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 28 KiB/s rd, 341 B/s wr, 43 op/s 2026-03-07T08:15:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:05 vm09 bash[21220]: cluster 2026-03-07T07:15:03.657956+0000 mgr.vm07.yrfcuj (mgr.14201) 1122 : cluster [DBG] pgmap v660: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 28 KiB/s rd, 341 B/s wr, 43 op/s 2026-03-07T08:15:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:05 vm09 bash[21220]: audit 2026-03-07T07:15:05.428984+0000 mon.vm07 (mon.0) 1272 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:05 vm09 bash[21220]: audit 2026-03-07T07:15:05.428984+0000 mon.vm07 (mon.0) 1272 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:05 vm09 bash[21220]: audit 2026-03-07T07:15:05.433422+0000 mon.vm07 (mon.0) 1273 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:05 vm09 bash[21220]: audit 2026-03-07T07:15:05.433422+0000 mon.vm07 (mon.0) 1273 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:05 vm09 bash[21220]: audit 2026-03-07T07:15:05.434206+0000 mon.vm07 (mon.0) 1274 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:05 vm09 bash[21220]: audit 2026-03-07T07:15:05.434206+0000 mon.vm07 (mon.0) 1274 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:05 vm09 bash[21220]: audit 2026-03-07T07:15:05.434670+0000 mon.vm07 (mon.0) 1275 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:05 vm09 bash[21220]: audit 2026-03-07T07:15:05.434670+0000 mon.vm07 (mon.0) 1275 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:05 vm09 bash[21220]: audit 2026-03-07T07:15:05.438176+0000 mon.vm07 (mon.0) 1276 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:05 vm09 bash[21220]: audit 2026-03-07T07:15:05.438176+0000 mon.vm07 (mon.0) 1276 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:05 vm09 bash[21220]: audit 2026-03-07T07:15:05.439662+0000 mon.vm07 (mon.0) 1277 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:05.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:05 vm09 bash[21220]: audit 2026-03-07T07:15:05.439662+0000 mon.vm07 (mon.0) 1277 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:05.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:05 vm07 bash[17031]: cluster 2026-03-07T07:15:03.657956+0000 mgr.vm07.yrfcuj (mgr.14201) 1122 : cluster [DBG] pgmap v660: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 28 KiB/s rd, 341 B/s wr, 43 op/s 2026-03-07T08:15:05.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:05 vm07 bash[17031]: cluster 2026-03-07T07:15:03.657956+0000 mgr.vm07.yrfcuj (mgr.14201) 1122 : cluster [DBG] pgmap v660: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 28 KiB/s rd, 341 B/s wr, 43 op/s 2026-03-07T08:15:05.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:05 vm07 bash[17031]: audit 2026-03-07T07:15:05.428984+0000 mon.vm07 (mon.0) 1272 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:05.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:05 vm07 bash[17031]: audit 2026-03-07T07:15:05.428984+0000 mon.vm07 (mon.0) 1272 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:05.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:05 vm07 bash[17031]: audit 2026-03-07T07:15:05.433422+0000 mon.vm07 (mon.0) 1273 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:05.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:05 vm07 bash[17031]: audit 2026-03-07T07:15:05.433422+0000 mon.vm07 (mon.0) 1273 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:05.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:05 vm07 bash[17031]: audit 2026-03-07T07:15:05.434206+0000 mon.vm07 (mon.0) 1274 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:05.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:05 vm07 bash[17031]: audit 2026-03-07T07:15:05.434206+0000 mon.vm07 (mon.0) 1274 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:05.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:05 vm07 bash[17031]: audit 2026-03-07T07:15:05.434670+0000 mon.vm07 (mon.0) 1275 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:05.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:05 vm07 bash[17031]: audit 2026-03-07T07:15:05.434670+0000 mon.vm07 (mon.0) 1275 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:05.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:05 vm07 bash[17031]: audit 2026-03-07T07:15:05.438176+0000 mon.vm07 (mon.0) 1276 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:05.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:05 vm07 bash[17031]: audit 2026-03-07T07:15:05.438176+0000 mon.vm07 (mon.0) 1276 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:05.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:05 vm07 bash[17031]: audit 2026-03-07T07:15:05.439662+0000 mon.vm07 (mon.0) 1277 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:05.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:05 vm07 bash[17031]: audit 2026-03-07T07:15:05.439662+0000 mon.vm07 (mon.0) 1277 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:06.855 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm07.rzyepz vm07 *:9000,9001 stopped 1s ago 17m - - 2026-03-07T08:15:06.859 INFO:teuthology.orchestra.run.vm07.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-07T08:15:06.859 INFO:teuthology.orchestra.run.vm07.stderr: Dload Upload Total Spent Left Speed 2026-03-07T08:15:06.860 INFO:teuthology.orchestra.run.vm07.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 182k 0 --:--:-- --:--:-- --:--:-- 182k 2026-03-07T08:15:07.031 INFO:teuthology.orchestra.run.vm07.stdout:anonymousScheduled to start haproxy.rgw.foo.vm07.rzyepz on host 'vm07' 2026-03-07T08:15:07.234 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for haproxy.rgw.foo.vm07.rzyepz to start 2026-03-07T08:15:07.402 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:15:07.402 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm07.rzyepz vm07 *:9000,9001 stopped 1s ago 17m - - 2026-03-07T08:15:07.402 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm09.hhbnqk vm09 *:9000,9001 running (17m) 15s ago 17m 3663k - 2.3.17-d1c9119 e85424b0d443 7325e33789e3 2026-03-07T08:15:07.610 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:15:08.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:07 vm09 bash[21220]: cluster 2026-03-07T07:15:05.658741+0000 mgr.vm07.yrfcuj (mgr.14201) 1123 : cluster [DBG] pgmap v661: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:08.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:07 vm09 bash[21220]: cluster 2026-03-07T07:15:05.658741+0000 mgr.vm07.yrfcuj (mgr.14201) 1123 : cluster [DBG] pgmap v661: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:08.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:07 vm09 bash[21220]: audit 2026-03-07T07:15:07.032049+0000 mon.vm07 (mon.0) 1278 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:08.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:07 vm09 bash[21220]: audit 2026-03-07T07:15:07.032049+0000 mon.vm07 (mon.0) 1278 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:08.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:07 vm09 bash[21220]: audit 2026-03-07T07:15:07.036483+0000 mon.vm07 (mon.0) 1279 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:08.108 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:07 vm09 bash[21220]: audit 2026-03-07T07:15:07.036483+0000 mon.vm07 (mon.0) 1279 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:08.108 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:07 vm09 bash[21220]: audit 2026-03-07T07:15:07.037270+0000 mon.vm07 (mon.0) 1280 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:08.108 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:07 vm09 bash[21220]: audit 2026-03-07T07:15:07.037270+0000 mon.vm07 (mon.0) 1280 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:08.108 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:07 vm09 bash[21220]: audit 2026-03-07T07:15:07.038613+0000 mon.vm07 (mon.0) 1281 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:08.108 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:07 vm09 bash[21220]: audit 2026-03-07T07:15:07.038613+0000 mon.vm07 (mon.0) 1281 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:08.108 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:07 vm09 bash[21220]: audit 2026-03-07T07:15:07.039061+0000 mon.vm07 (mon.0) 1282 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:08.108 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:07 vm09 bash[21220]: audit 2026-03-07T07:15:07.039061+0000 mon.vm07 (mon.0) 1282 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:08.108 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:07 vm09 bash[21220]: audit 2026-03-07T07:15:07.043463+0000 mon.vm07 (mon.0) 1283 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:08.108 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:07 vm09 bash[21220]: audit 2026-03-07T07:15:07.043463+0000 mon.vm07 (mon.0) 1283 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:08.108 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:07 vm09 bash[21220]: audit 2026-03-07T07:15:07.044795+0000 mon.vm07 (mon.0) 1284 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:08.108 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:07 vm09 bash[21220]: audit 2026-03-07T07:15:07.044795+0000 mon.vm07 (mon.0) 1284 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:08.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:07 vm07 bash[17031]: cluster 2026-03-07T07:15:05.658741+0000 mgr.vm07.yrfcuj (mgr.14201) 1123 : cluster [DBG] pgmap v661: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:08.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:07 vm07 bash[17031]: cluster 2026-03-07T07:15:05.658741+0000 mgr.vm07.yrfcuj (mgr.14201) 1123 : cluster [DBG] pgmap v661: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:08.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:07 vm07 bash[17031]: audit 2026-03-07T07:15:07.032049+0000 mon.vm07 (mon.0) 1278 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:08.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:07 vm07 bash[17031]: audit 2026-03-07T07:15:07.032049+0000 mon.vm07 (mon.0) 1278 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:08.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:07 vm07 bash[17031]: audit 2026-03-07T07:15:07.036483+0000 mon.vm07 (mon.0) 1279 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:08.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:07 vm07 bash[17031]: audit 2026-03-07T07:15:07.036483+0000 mon.vm07 (mon.0) 1279 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:08.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:07 vm07 bash[17031]: audit 2026-03-07T07:15:07.037270+0000 mon.vm07 (mon.0) 1280 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:08.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:07 vm07 bash[17031]: audit 2026-03-07T07:15:07.037270+0000 mon.vm07 (mon.0) 1280 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:08.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:07 vm07 bash[17031]: audit 2026-03-07T07:15:07.038613+0000 mon.vm07 (mon.0) 1281 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:08.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:07 vm07 bash[17031]: audit 2026-03-07T07:15:07.038613+0000 mon.vm07 (mon.0) 1281 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:08.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:07 vm07 bash[17031]: audit 2026-03-07T07:15:07.039061+0000 mon.vm07 (mon.0) 1282 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:08.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:07 vm07 bash[17031]: audit 2026-03-07T07:15:07.039061+0000 mon.vm07 (mon.0) 1282 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:08.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:07 vm07 bash[17031]: audit 2026-03-07T07:15:07.043463+0000 mon.vm07 (mon.0) 1283 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:08.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:07 vm07 bash[17031]: audit 2026-03-07T07:15:07.043463+0000 mon.vm07 (mon.0) 1283 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:08.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:07 vm07 bash[17031]: audit 2026-03-07T07:15:07.044795+0000 mon.vm07 (mon.0) 1284 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:08.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:07 vm07 bash[17031]: audit 2026-03-07T07:15:07.044795+0000 mon.vm07 (mon.0) 1284 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:09.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:08 vm09 bash[21220]: audit 2026-03-07T07:15:06.845421+0000 mgr.vm07.yrfcuj (mgr.14201) 1124 : audit [DBG] from='client.16918 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:09.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:08 vm09 bash[21220]: audit 2026-03-07T07:15:06.845421+0000 mgr.vm07.yrfcuj (mgr.14201) 1124 : audit [DBG] from='client.16918 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:09.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:08 vm09 bash[21220]: audit 2026-03-07T07:15:07.026499+0000 mgr.vm07.yrfcuj (mgr.14201) 1125 : audit [DBG] from='client.16922 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "haproxy.rgw.foo.vm07.rzyepz", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:09.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:08 vm09 bash[21220]: audit 2026-03-07T07:15:07.026499+0000 mgr.vm07.yrfcuj (mgr.14201) 1125 : audit [DBG] from='client.16922 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "haproxy.rgw.foo.vm07.rzyepz", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:09.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:08 vm09 bash[21220]: cephadm 2026-03-07T07:15:07.026834+0000 mgr.vm07.yrfcuj (mgr.14201) 1126 : cephadm [INF] Schedule start daemon haproxy.rgw.foo.vm07.rzyepz 2026-03-07T08:15:09.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:08 vm09 bash[21220]: cephadm 2026-03-07T07:15:07.026834+0000 mgr.vm07.yrfcuj (mgr.14201) 1126 : cephadm [INF] Schedule start daemon haproxy.rgw.foo.vm07.rzyepz 2026-03-07T08:15:09.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:08 vm09 bash[21220]: audit 2026-03-07T07:15:07.224459+0000 mgr.vm07.yrfcuj (mgr.14201) 1127 : audit [DBG] from='client.16926 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:09.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:08 vm09 bash[21220]: audit 2026-03-07T07:15:07.224459+0000 mgr.vm07.yrfcuj (mgr.14201) 1127 : audit [DBG] from='client.16926 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:09.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:08 vm09 bash[21220]: audit 2026-03-07T07:15:07.405526+0000 mgr.vm07.yrfcuj (mgr.14201) 1128 : audit [DBG] from='client.16930 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:09.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:08 vm09 bash[21220]: audit 2026-03-07T07:15:07.405526+0000 mgr.vm07.yrfcuj (mgr.14201) 1128 : audit [DBG] from='client.16930 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:09.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:08 vm09 bash[21220]: audit 2026-03-07T07:15:07.615944+0000 mon.vm07 (mon.0) 1285 : audit [DBG] from='client.? 192.168.123.107:0/2730780024' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:09.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:08 vm09 bash[21220]: audit 2026-03-07T07:15:07.615944+0000 mon.vm07 (mon.0) 1285 : audit [DBG] from='client.? 192.168.123.107:0/2730780024' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:09.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:08 vm09 bash[21220]: cluster 2026-03-07T07:15:07.659225+0000 mgr.vm07.yrfcuj (mgr.14201) 1129 : cluster [DBG] pgmap v662: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:09.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:08 vm09 bash[21220]: cluster 2026-03-07T07:15:07.659225+0000 mgr.vm07.yrfcuj (mgr.14201) 1129 : cluster [DBG] pgmap v662: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:09.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:08 vm07 bash[17031]: audit 2026-03-07T07:15:06.845421+0000 mgr.vm07.yrfcuj (mgr.14201) 1124 : audit [DBG] from='client.16918 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:09.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:08 vm07 bash[17031]: audit 2026-03-07T07:15:06.845421+0000 mgr.vm07.yrfcuj (mgr.14201) 1124 : audit [DBG] from='client.16918 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:09.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:08 vm07 bash[17031]: audit 2026-03-07T07:15:07.026499+0000 mgr.vm07.yrfcuj (mgr.14201) 1125 : audit [DBG] from='client.16922 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "haproxy.rgw.foo.vm07.rzyepz", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:09.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:08 vm07 bash[17031]: audit 2026-03-07T07:15:07.026499+0000 mgr.vm07.yrfcuj (mgr.14201) 1125 : audit [DBG] from='client.16922 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "haproxy.rgw.foo.vm07.rzyepz", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:09.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:08 vm07 bash[17031]: cephadm 2026-03-07T07:15:07.026834+0000 mgr.vm07.yrfcuj (mgr.14201) 1126 : cephadm [INF] Schedule start daemon haproxy.rgw.foo.vm07.rzyepz 2026-03-07T08:15:09.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:08 vm07 bash[17031]: cephadm 2026-03-07T07:15:07.026834+0000 mgr.vm07.yrfcuj (mgr.14201) 1126 : cephadm [INF] Schedule start daemon haproxy.rgw.foo.vm07.rzyepz 2026-03-07T08:15:09.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:08 vm07 bash[17031]: audit 2026-03-07T07:15:07.224459+0000 mgr.vm07.yrfcuj (mgr.14201) 1127 : audit [DBG] from='client.16926 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:09.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:08 vm07 bash[17031]: audit 2026-03-07T07:15:07.224459+0000 mgr.vm07.yrfcuj (mgr.14201) 1127 : audit [DBG] from='client.16926 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:09.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:08 vm07 bash[17031]: audit 2026-03-07T07:15:07.405526+0000 mgr.vm07.yrfcuj (mgr.14201) 1128 : audit [DBG] from='client.16930 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:09.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:08 vm07 bash[17031]: audit 2026-03-07T07:15:07.405526+0000 mgr.vm07.yrfcuj (mgr.14201) 1128 : audit [DBG] from='client.16930 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:09.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:08 vm07 bash[17031]: audit 2026-03-07T07:15:07.615944+0000 mon.vm07 (mon.0) 1285 : audit [DBG] from='client.? 192.168.123.107:0/2730780024' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:09.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:08 vm07 bash[17031]: audit 2026-03-07T07:15:07.615944+0000 mon.vm07 (mon.0) 1285 : audit [DBG] from='client.? 192.168.123.107:0/2730780024' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:09.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:08 vm07 bash[17031]: cluster 2026-03-07T07:15:07.659225+0000 mgr.vm07.yrfcuj (mgr.14201) 1129 : cluster [DBG] pgmap v662: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:09.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:08 vm07 bash[17031]: cluster 2026-03-07T07:15:07.659225+0000 mgr.vm07.yrfcuj (mgr.14201) 1129 : cluster [DBG] pgmap v662: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:11.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:10 vm09 bash[21220]: cluster 2026-03-07T07:15:09.659602+0000 mgr.vm07.yrfcuj (mgr.14201) 1130 : cluster [DBG] pgmap v663: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:11.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:10 vm09 bash[21220]: cluster 2026-03-07T07:15:09.659602+0000 mgr.vm07.yrfcuj (mgr.14201) 1130 : cluster [DBG] pgmap v663: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:11.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:10 vm09 bash[21220]: audit 2026-03-07T07:15:10.148011+0000 mon.vm07 (mon.0) 1286 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:15:11.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:10 vm09 bash[21220]: audit 2026-03-07T07:15:10.148011+0000 mon.vm07 (mon.0) 1286 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:15:11.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:10 vm07 bash[17031]: cluster 2026-03-07T07:15:09.659602+0000 mgr.vm07.yrfcuj (mgr.14201) 1130 : cluster [DBG] pgmap v663: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:11.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:10 vm07 bash[17031]: cluster 2026-03-07T07:15:09.659602+0000 mgr.vm07.yrfcuj (mgr.14201) 1130 : cluster [DBG] pgmap v663: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:11.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:10 vm07 bash[17031]: audit 2026-03-07T07:15:10.148011+0000 mon.vm07 (mon.0) 1286 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:15:11.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:10 vm07 bash[17031]: audit 2026-03-07T07:15:10.148011+0000 mon.vm07 (mon.0) 1286 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:15:12.787 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for haproxy.rgw.foo.vm07.rzyepz to start 2026-03-07T08:15:12.942 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:15:12.942 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm07.rzyepz vm07 *:9000,9001 stopped 7s ago 17m - - 2026-03-07T08:15:12.942 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm09.hhbnqk vm09 *:9000,9001 running (17m) 21s ago 17m 3663k - 2.3.17-d1c9119 e85424b0d443 7325e33789e3 2026-03-07T08:15:13.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:12 vm09 bash[21220]: cluster 2026-03-07T07:15:11.660043+0000 mgr.vm07.yrfcuj (mgr.14201) 1131 : cluster [DBG] pgmap v664: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:13.107 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:12 vm09 bash[21220]: cluster 2026-03-07T07:15:11.660043+0000 mgr.vm07.yrfcuj (mgr.14201) 1131 : cluster [DBG] pgmap v664: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:13.131 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:15:13.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:12 vm07 bash[17031]: cluster 2026-03-07T07:15:11.660043+0000 mgr.vm07.yrfcuj (mgr.14201) 1131 : cluster [DBG] pgmap v664: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:13.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:12 vm07 bash[17031]: cluster 2026-03-07T07:15:11.660043+0000 mgr.vm07.yrfcuj (mgr.14201) 1131 : cluster [DBG] pgmap v664: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:14.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:13 vm09 bash[21220]: audit 2026-03-07T07:15:12.780955+0000 mgr.vm07.yrfcuj (mgr.14201) 1132 : audit [DBG] from='client.16938 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:14.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:13 vm09 bash[21220]: audit 2026-03-07T07:15:12.780955+0000 mgr.vm07.yrfcuj (mgr.14201) 1132 : audit [DBG] from='client.16938 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:14.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:13 vm09 bash[21220]: audit 2026-03-07T07:15:12.945380+0000 mgr.vm07.yrfcuj (mgr.14201) 1133 : audit [DBG] from='client.16942 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:14.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:13 vm09 bash[21220]: audit 2026-03-07T07:15:12.945380+0000 mgr.vm07.yrfcuj (mgr.14201) 1133 : audit [DBG] from='client.16942 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:14.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:13 vm09 bash[21220]: audit 2026-03-07T07:15:13.130897+0000 mon.vm09 (mon.1) 53 : audit [DBG] from='client.? 192.168.123.107:0/1676399327' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:14.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:13 vm09 bash[21220]: audit 2026-03-07T07:15:13.130897+0000 mon.vm09 (mon.1) 53 : audit [DBG] from='client.? 192.168.123.107:0/1676399327' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:14.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:13 vm07 bash[17031]: audit 2026-03-07T07:15:12.780955+0000 mgr.vm07.yrfcuj (mgr.14201) 1132 : audit [DBG] from='client.16938 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:14.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:13 vm07 bash[17031]: audit 2026-03-07T07:15:12.780955+0000 mgr.vm07.yrfcuj (mgr.14201) 1132 : audit [DBG] from='client.16938 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:14.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:13 vm07 bash[17031]: audit 2026-03-07T07:15:12.945380+0000 mgr.vm07.yrfcuj (mgr.14201) 1133 : audit [DBG] from='client.16942 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:14.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:13 vm07 bash[17031]: audit 2026-03-07T07:15:12.945380+0000 mgr.vm07.yrfcuj (mgr.14201) 1133 : audit [DBG] from='client.16942 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:14.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:13 vm07 bash[17031]: audit 2026-03-07T07:15:13.130897+0000 mon.vm09 (mon.1) 53 : audit [DBG] from='client.? 192.168.123.107:0/1676399327' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:14.147 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:13 vm07 bash[17031]: audit 2026-03-07T07:15:13.130897+0000 mon.vm09 (mon.1) 53 : audit [DBG] from='client.? 192.168.123.107:0/1676399327' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:14.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:14 vm07 bash[17031]: cluster 2026-03-07T07:15:13.660575+0000 mgr.vm07.yrfcuj (mgr.14201) 1134 : cluster [DBG] pgmap v665: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:14.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:14 vm07 bash[17031]: cluster 2026-03-07T07:15:13.660575+0000 mgr.vm07.yrfcuj (mgr.14201) 1134 : cluster [DBG] pgmap v665: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:15.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:14 vm09 bash[21220]: cluster 2026-03-07T07:15:13.660575+0000 mgr.vm07.yrfcuj (mgr.14201) 1134 : cluster [DBG] pgmap v665: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:15.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:14 vm09 bash[21220]: cluster 2026-03-07T07:15:13.660575+0000 mgr.vm07.yrfcuj (mgr.14201) 1134 : cluster [DBG] pgmap v665: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:17.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:16 vm07 bash[17031]: cluster 2026-03-07T07:15:15.661086+0000 mgr.vm07.yrfcuj (mgr.14201) 1135 : cluster [DBG] pgmap v666: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:17.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:16 vm07 bash[17031]: cluster 2026-03-07T07:15:15.661086+0000 mgr.vm07.yrfcuj (mgr.14201) 1135 : cluster [DBG] pgmap v666: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:17.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:16 vm07 bash[17031]: audit 2026-03-07T07:15:15.964429+0000 mon.vm07 (mon.0) 1287 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:17.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:16 vm07 bash[17031]: audit 2026-03-07T07:15:15.964429+0000 mon.vm07 (mon.0) 1287 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:17.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:16 vm07 bash[17031]: audit 2026-03-07T07:15:15.984849+0000 mon.vm07 (mon.0) 1288 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:17.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:16 vm07 bash[17031]: audit 2026-03-07T07:15:15.984849+0000 mon.vm07 (mon.0) 1288 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:17.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:16 vm07 bash[17031]: audit 2026-03-07T07:15:16.036295+0000 mon.vm07 (mon.0) 1289 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:17.328 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:16 vm07 bash[17031]: audit 2026-03-07T07:15:16.036295+0000 mon.vm07 (mon.0) 1289 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:17.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:16 vm09 bash[21220]: cluster 2026-03-07T07:15:15.661086+0000 mgr.vm07.yrfcuj (mgr.14201) 1135 : cluster [DBG] pgmap v666: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:17.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:16 vm09 bash[21220]: cluster 2026-03-07T07:15:15.661086+0000 mgr.vm07.yrfcuj (mgr.14201) 1135 : cluster [DBG] pgmap v666: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:17.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:16 vm09 bash[21220]: audit 2026-03-07T07:15:15.964429+0000 mon.vm07 (mon.0) 1287 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:17.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:16 vm09 bash[21220]: audit 2026-03-07T07:15:15.964429+0000 mon.vm07 (mon.0) 1287 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:17.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:16 vm09 bash[21220]: audit 2026-03-07T07:15:15.984849+0000 mon.vm07 (mon.0) 1288 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:17.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:16 vm09 bash[21220]: audit 2026-03-07T07:15:15.984849+0000 mon.vm07 (mon.0) 1288 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:17.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:16 vm09 bash[21220]: audit 2026-03-07T07:15:16.036295+0000 mon.vm07 (mon.0) 1289 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:17.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:16 vm09 bash[21220]: audit 2026-03-07T07:15:16.036295+0000 mon.vm07 (mon.0) 1289 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:18.317 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for haproxy.rgw.foo.vm07.rzyepz to start 2026-03-07T08:15:18.485 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:15:18.485 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm07.rzyepz vm07 *:9000,9001 stopped 13s ago 17m - - 2026-03-07T08:15:18.485 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm09.hhbnqk vm09 *:9000,9001 running (17m) 26s ago 17m 3663k - 2.3.17-d1c9119 e85424b0d443 7325e33789e3 2026-03-07T08:15:18.688 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:15:19.327 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:19 vm07 bash[17031]: cluster 2026-03-07T07:15:17.661564+0000 mgr.vm07.yrfcuj (mgr.14201) 1136 : cluster [DBG] pgmap v667: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:19.327 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:19 vm07 bash[17031]: cluster 2026-03-07T07:15:17.661564+0000 mgr.vm07.yrfcuj (mgr.14201) 1136 : cluster [DBG] pgmap v667: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:19.327 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:19 vm07 bash[17031]: audit 2026-03-07T07:15:18.306909+0000 mgr.vm07.yrfcuj (mgr.14201) 1137 : audit [DBG] from='client.16950 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:19.327 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:19 vm07 bash[17031]: audit 2026-03-07T07:15:18.306909+0000 mgr.vm07.yrfcuj (mgr.14201) 1137 : audit [DBG] from='client.16950 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:19.327 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:19 vm07 bash[17031]: audit 2026-03-07T07:15:18.488711+0000 mgr.vm07.yrfcuj (mgr.14201) 1138 : audit [DBG] from='client.16954 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:19.327 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:19 vm07 bash[17031]: audit 2026-03-07T07:15:18.488711+0000 mgr.vm07.yrfcuj (mgr.14201) 1138 : audit [DBG] from='client.16954 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:19.327 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:19 vm07 bash[17031]: audit 2026-03-07T07:15:18.694631+0000 mon.vm07 (mon.0) 1290 : audit [DBG] from='client.? 192.168.123.107:0/303502029' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:19.327 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:19 vm07 bash[17031]: audit 2026-03-07T07:15:18.694631+0000 mon.vm07 (mon.0) 1290 : audit [DBG] from='client.? 192.168.123.107:0/303502029' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:19.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:19 vm09 bash[21220]: cluster 2026-03-07T07:15:17.661564+0000 mgr.vm07.yrfcuj (mgr.14201) 1136 : cluster [DBG] pgmap v667: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:19.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:19 vm09 bash[21220]: cluster 2026-03-07T07:15:17.661564+0000 mgr.vm07.yrfcuj (mgr.14201) 1136 : cluster [DBG] pgmap v667: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:19.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:19 vm09 bash[21220]: audit 2026-03-07T07:15:18.306909+0000 mgr.vm07.yrfcuj (mgr.14201) 1137 : audit [DBG] from='client.16950 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:19.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:19 vm09 bash[21220]: audit 2026-03-07T07:15:18.306909+0000 mgr.vm07.yrfcuj (mgr.14201) 1137 : audit [DBG] from='client.16950 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:19.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:19 vm09 bash[21220]: audit 2026-03-07T07:15:18.488711+0000 mgr.vm07.yrfcuj (mgr.14201) 1138 : audit [DBG] from='client.16954 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:19.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:19 vm09 bash[21220]: audit 2026-03-07T07:15:18.488711+0000 mgr.vm07.yrfcuj (mgr.14201) 1138 : audit [DBG] from='client.16954 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:19.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:19 vm09 bash[21220]: audit 2026-03-07T07:15:18.694631+0000 mon.vm07 (mon.0) 1290 : audit [DBG] from='client.? 192.168.123.107:0/303502029' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:19.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:19 vm09 bash[21220]: audit 2026-03-07T07:15:18.694631+0000 mon.vm07 (mon.0) 1290 : audit [DBG] from='client.? 192.168.123.107:0/303502029' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:21.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:21 vm09 bash[21220]: cluster 2026-03-07T07:15:19.661879+0000 mgr.vm07.yrfcuj (mgr.14201) 1139 : cluster [DBG] pgmap v668: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:21.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:21 vm09 bash[21220]: cluster 2026-03-07T07:15:19.661879+0000 mgr.vm07.yrfcuj (mgr.14201) 1139 : cluster [DBG] pgmap v668: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:21.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:21 vm09 bash[21220]: audit 2026-03-07T07:15:20.714222+0000 mon.vm07 (mon.0) 1291 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:21.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:21 vm09 bash[21220]: audit 2026-03-07T07:15:20.714222+0000 mon.vm07 (mon.0) 1291 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:21.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:21 vm09 bash[21220]: audit 2026-03-07T07:15:20.719482+0000 mon.vm07 (mon.0) 1292 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:21.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:21 vm09 bash[21220]: audit 2026-03-07T07:15:20.719482+0000 mon.vm07 (mon.0) 1292 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:21.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:21 vm09 bash[21220]: audit 2026-03-07T07:15:20.720425+0000 mon.vm07 (mon.0) 1293 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:21.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:21 vm09 bash[21220]: audit 2026-03-07T07:15:20.720425+0000 mon.vm07 (mon.0) 1293 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:21.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:21 vm09 bash[21220]: audit 2026-03-07T07:15:20.721004+0000 mon.vm07 (mon.0) 1294 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:21.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:21 vm09 bash[21220]: audit 2026-03-07T07:15:20.721004+0000 mon.vm07 (mon.0) 1294 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:21.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:21 vm09 bash[21220]: audit 2026-03-07T07:15:20.725107+0000 mon.vm07 (mon.0) 1295 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:21.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:21 vm09 bash[21220]: audit 2026-03-07T07:15:20.725107+0000 mon.vm07 (mon.0) 1295 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:21.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:21 vm09 bash[21220]: audit 2026-03-07T07:15:20.726713+0000 mon.vm07 (mon.0) 1296 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:21.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:21 vm09 bash[21220]: audit 2026-03-07T07:15:20.726713+0000 mon.vm07 (mon.0) 1296 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:21.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:21 vm07 bash[17031]: cluster 2026-03-07T07:15:19.661879+0000 mgr.vm07.yrfcuj (mgr.14201) 1139 : cluster [DBG] pgmap v668: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:21.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:21 vm07 bash[17031]: cluster 2026-03-07T07:15:19.661879+0000 mgr.vm07.yrfcuj (mgr.14201) 1139 : cluster [DBG] pgmap v668: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:21.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:21 vm07 bash[17031]: audit 2026-03-07T07:15:20.714222+0000 mon.vm07 (mon.0) 1291 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:21.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:21 vm07 bash[17031]: audit 2026-03-07T07:15:20.714222+0000 mon.vm07 (mon.0) 1291 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:21.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:21 vm07 bash[17031]: audit 2026-03-07T07:15:20.719482+0000 mon.vm07 (mon.0) 1292 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:21.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:21 vm07 bash[17031]: audit 2026-03-07T07:15:20.719482+0000 mon.vm07 (mon.0) 1292 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:21.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:21 vm07 bash[17031]: audit 2026-03-07T07:15:20.720425+0000 mon.vm07 (mon.0) 1293 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:21.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:21 vm07 bash[17031]: audit 2026-03-07T07:15:20.720425+0000 mon.vm07 (mon.0) 1293 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:21.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:21 vm07 bash[17031]: audit 2026-03-07T07:15:20.721004+0000 mon.vm07 (mon.0) 1294 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:21.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:21 vm07 bash[17031]: audit 2026-03-07T07:15:20.721004+0000 mon.vm07 (mon.0) 1294 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:21.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:21 vm07 bash[17031]: audit 2026-03-07T07:15:20.725107+0000 mon.vm07 (mon.0) 1295 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:21.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:21 vm07 bash[17031]: audit 2026-03-07T07:15:20.725107+0000 mon.vm07 (mon.0) 1295 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:21.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:21 vm07 bash[17031]: audit 2026-03-07T07:15:20.726713+0000 mon.vm07 (mon.0) 1296 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:21.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:21 vm07 bash[17031]: audit 2026-03-07T07:15:20.726713+0000 mon.vm07 (mon.0) 1296 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:23.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:23 vm09 bash[21220]: cluster 2026-03-07T07:15:21.662335+0000 mgr.vm07.yrfcuj (mgr.14201) 1140 : cluster [DBG] pgmap v669: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:23.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:23 vm09 bash[21220]: cluster 2026-03-07T07:15:21.662335+0000 mgr.vm07.yrfcuj (mgr.14201) 1140 : cluster [DBG] pgmap v669: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:23.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:23 vm07 bash[17031]: cluster 2026-03-07T07:15:21.662335+0000 mgr.vm07.yrfcuj (mgr.14201) 1140 : cluster [DBG] pgmap v669: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:23.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:23 vm07 bash[17031]: cluster 2026-03-07T07:15:21.662335+0000 mgr.vm07.yrfcuj (mgr.14201) 1140 : cluster [DBG] pgmap v669: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:23.869 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm07.rzyepz vm07 *:9000,9001 running (7s) 3s ago 17m 3516k - 2.3.17-d1c9119 e85424b0d443 5def5ff2697f 2026-03-07T08:15:24.035 INFO:teuthology.orchestra.run.vm07.stdout:Scheduled to stop haproxy.rgw.foo.vm09.hhbnqk on host 'vm09' 2026-03-07T08:15:24.235 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for haproxy.rgw.foo.vm09.hhbnqk to stop 2026-03-07T08:15:24.389 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:15:24.389 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm07.rzyepz vm07 *:9000,9001 running (8s) 3s ago 17m 3516k - 2.3.17-d1c9119 e85424b0d443 5def5ff2697f 2026-03-07T08:15:24.389 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm09.hhbnqk vm09 *:9000,9001 running (17m) 32s ago 17m 3663k - 2.3.17-d1c9119 e85424b0d443 7325e33789e3 2026-03-07T08:15:24.585 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: cluster 2026-03-07T07:15:23.662748+0000 mgr.vm07.yrfcuj (mgr.14201) 1141 : cluster [DBG] pgmap v670: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: cluster 2026-03-07T07:15:23.662748+0000 mgr.vm07.yrfcuj (mgr.14201) 1141 : cluster [DBG] pgmap v670: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:23.860834+0000 mgr.vm07.yrfcuj (mgr.14201) 1142 : audit [DBG] from='client.16962 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:23.860834+0000 mgr.vm07.yrfcuj (mgr.14201) 1142 : audit [DBG] from='client.16962 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.030311+0000 mgr.vm07.yrfcuj (mgr.14201) 1143 : audit [DBG] from='client.16966 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "haproxy.rgw.foo.vm09.hhbnqk", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.030311+0000 mgr.vm07.yrfcuj (mgr.14201) 1143 : audit [DBG] from='client.16966 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "haproxy.rgw.foo.vm09.hhbnqk", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: cephadm 2026-03-07T07:15:24.030668+0000 mgr.vm07.yrfcuj (mgr.14201) 1144 : cephadm [INF] Schedule stop daemon haproxy.rgw.foo.vm09.hhbnqk 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: cephadm 2026-03-07T07:15:24.030668+0000 mgr.vm07.yrfcuj (mgr.14201) 1144 : cephadm [INF] Schedule stop daemon haproxy.rgw.foo.vm09.hhbnqk 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.035619+0000 mon.vm07 (mon.0) 1297 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.035619+0000 mon.vm07 (mon.0) 1297 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.040016+0000 mon.vm07 (mon.0) 1298 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.040016+0000 mon.vm07 (mon.0) 1298 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.040898+0000 mon.vm07 (mon.0) 1299 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.040898+0000 mon.vm07 (mon.0) 1299 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.042064+0000 mon.vm07 (mon.0) 1300 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.042064+0000 mon.vm07 (mon.0) 1300 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.042426+0000 mon.vm07 (mon.0) 1301 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.042426+0000 mon.vm07 (mon.0) 1301 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.046735+0000 mon.vm07 (mon.0) 1302 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.046735+0000 mon.vm07 (mon.0) 1302 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.048483+0000 mon.vm07 (mon.0) 1303 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.048483+0000 mon.vm07 (mon.0) 1303 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.223572+0000 mgr.vm07.yrfcuj (mgr.14201) 1145 : audit [DBG] from='client.16970 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.223572+0000 mgr.vm07.yrfcuj (mgr.14201) 1145 : audit [DBG] from='client.16970 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.393189+0000 mgr.vm07.yrfcuj (mgr.14201) 1146 : audit [DBG] from='client.16974 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.393189+0000 mgr.vm07.yrfcuj (mgr.14201) 1146 : audit [DBG] from='client.16974 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.591128+0000 mon.vm07 (mon.0) 1304 : audit [DBG] from='client.? 192.168.123.107:0/3012128028' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:25.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:25 vm09 bash[21220]: audit 2026-03-07T07:15:24.591128+0000 mon.vm07 (mon.0) 1304 : audit [DBG] from='client.? 192.168.123.107:0/3012128028' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:25.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: cluster 2026-03-07T07:15:23.662748+0000 mgr.vm07.yrfcuj (mgr.14201) 1141 : cluster [DBG] pgmap v670: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:25.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: cluster 2026-03-07T07:15:23.662748+0000 mgr.vm07.yrfcuj (mgr.14201) 1141 : cluster [DBG] pgmap v670: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:25.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:23.860834+0000 mgr.vm07.yrfcuj (mgr.14201) 1142 : audit [DBG] from='client.16962 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:25.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:23.860834+0000 mgr.vm07.yrfcuj (mgr.14201) 1142 : audit [DBG] from='client.16962 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:25.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.030311+0000 mgr.vm07.yrfcuj (mgr.14201) 1143 : audit [DBG] from='client.16966 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "haproxy.rgw.foo.vm09.hhbnqk", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.030311+0000 mgr.vm07.yrfcuj (mgr.14201) 1143 : audit [DBG] from='client.16966 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "stop", "name": "haproxy.rgw.foo.vm09.hhbnqk", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: cephadm 2026-03-07T07:15:24.030668+0000 mgr.vm07.yrfcuj (mgr.14201) 1144 : cephadm [INF] Schedule stop daemon haproxy.rgw.foo.vm09.hhbnqk 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: cephadm 2026-03-07T07:15:24.030668+0000 mgr.vm07.yrfcuj (mgr.14201) 1144 : cephadm [INF] Schedule stop daemon haproxy.rgw.foo.vm09.hhbnqk 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.035619+0000 mon.vm07 (mon.0) 1297 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.035619+0000 mon.vm07 (mon.0) 1297 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.040016+0000 mon.vm07 (mon.0) 1298 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.040016+0000 mon.vm07 (mon.0) 1298 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.040898+0000 mon.vm07 (mon.0) 1299 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.040898+0000 mon.vm07 (mon.0) 1299 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.042064+0000 mon.vm07 (mon.0) 1300 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.042064+0000 mon.vm07 (mon.0) 1300 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.042426+0000 mon.vm07 (mon.0) 1301 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.042426+0000 mon.vm07 (mon.0) 1301 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.046735+0000 mon.vm07 (mon.0) 1302 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.046735+0000 mon.vm07 (mon.0) 1302 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.048483+0000 mon.vm07 (mon.0) 1303 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.048483+0000 mon.vm07 (mon.0) 1303 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.223572+0000 mgr.vm07.yrfcuj (mgr.14201) 1145 : audit [DBG] from='client.16970 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.223572+0000 mgr.vm07.yrfcuj (mgr.14201) 1145 : audit [DBG] from='client.16970 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.393189+0000 mgr.vm07.yrfcuj (mgr.14201) 1146 : audit [DBG] from='client.16974 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.393189+0000 mgr.vm07.yrfcuj (mgr.14201) 1146 : audit [DBG] from='client.16974 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.591128+0000 mon.vm07 (mon.0) 1304 : audit [DBG] from='client.? 192.168.123.107:0/3012128028' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:25.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:25 vm07 bash[17031]: audit 2026-03-07T07:15:24.591128+0000 mon.vm07 (mon.0) 1304 : audit [DBG] from='client.? 192.168.123.107:0/3012128028' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:26.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:26 vm09 bash[21220]: audit 2026-03-07T07:15:25.148108+0000 mon.vm07 (mon.0) 1305 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:15:26.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:26 vm09 bash[21220]: audit 2026-03-07T07:15:25.148108+0000 mon.vm07 (mon.0) 1305 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:15:26.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:26 vm07 bash[17031]: audit 2026-03-07T07:15:25.148108+0000 mon.vm07 (mon.0) 1305 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:15:26.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:26 vm07 bash[17031]: audit 2026-03-07T07:15:25.148108+0000 mon.vm07 (mon.0) 1305 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:15:27.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:27 vm09 bash[21220]: cluster 2026-03-07T07:15:25.663267+0000 mgr.vm07.yrfcuj (mgr.14201) 1147 : cluster [DBG] pgmap v671: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:27.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:27 vm09 bash[21220]: cluster 2026-03-07T07:15:25.663267+0000 mgr.vm07.yrfcuj (mgr.14201) 1147 : cluster [DBG] pgmap v671: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:27.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:27 vm07 bash[17031]: cluster 2026-03-07T07:15:25.663267+0000 mgr.vm07.yrfcuj (mgr.14201) 1147 : cluster [DBG] pgmap v671: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:27.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:27 vm07 bash[17031]: cluster 2026-03-07T07:15:25.663267+0000 mgr.vm07.yrfcuj (mgr.14201) 1147 : cluster [DBG] pgmap v671: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:29 vm09 bash[21220]: cluster 2026-03-07T07:15:27.663774+0000 mgr.vm07.yrfcuj (mgr.14201) 1148 : cluster [DBG] pgmap v672: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:29 vm09 bash[21220]: cluster 2026-03-07T07:15:27.663774+0000 mgr.vm07.yrfcuj (mgr.14201) 1148 : cluster [DBG] pgmap v672: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:29 vm09 bash[21220]: audit 2026-03-07T07:15:28.992298+0000 mon.vm07 (mon.0) 1306 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:29 vm09 bash[21220]: audit 2026-03-07T07:15:28.992298+0000 mon.vm07 (mon.0) 1306 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:29 vm09 bash[21220]: audit 2026-03-07T07:15:28.997677+0000 mon.vm07 (mon.0) 1307 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:29 vm09 bash[21220]: audit 2026-03-07T07:15:28.997677+0000 mon.vm07 (mon.0) 1307 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:29 vm09 bash[21220]: audit 2026-03-07T07:15:29.026314+0000 mon.vm07 (mon.0) 1308 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:29.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:29 vm09 bash[21220]: audit 2026-03-07T07:15:29.026314+0000 mon.vm07 (mon.0) 1308 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:29.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:29 vm07 bash[17031]: cluster 2026-03-07T07:15:27.663774+0000 mgr.vm07.yrfcuj (mgr.14201) 1148 : cluster [DBG] pgmap v672: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:29.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:29 vm07 bash[17031]: cluster 2026-03-07T07:15:27.663774+0000 mgr.vm07.yrfcuj (mgr.14201) 1148 : cluster [DBG] pgmap v672: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:29.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:29 vm07 bash[17031]: audit 2026-03-07T07:15:28.992298+0000 mon.vm07 (mon.0) 1306 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:29.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:29 vm07 bash[17031]: audit 2026-03-07T07:15:28.992298+0000 mon.vm07 (mon.0) 1306 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:29.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:29 vm07 bash[17031]: audit 2026-03-07T07:15:28.997677+0000 mon.vm07 (mon.0) 1307 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:29.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:29 vm07 bash[17031]: audit 2026-03-07T07:15:28.997677+0000 mon.vm07 (mon.0) 1307 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:29.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:29 vm07 bash[17031]: audit 2026-03-07T07:15:29.026314+0000 mon.vm07 (mon.0) 1308 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:29.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:29 vm07 bash[17031]: audit 2026-03-07T07:15:29.026314+0000 mon.vm07 (mon.0) 1308 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:29.758 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for haproxy.rgw.foo.vm09.hhbnqk to stop 2026-03-07T08:15:29.920 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:15:29.920 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm07.rzyepz vm07 *:9000,9001 running (13s) 9s ago 17m 3516k - 2.3.17-d1c9119 e85424b0d443 5def5ff2697f 2026-03-07T08:15:29.920 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm09.hhbnqk vm09 *:9000,9001 running (17m) 38s ago 17m 3663k - 2.3.17-d1c9119 e85424b0d443 7325e33789e3 2026-03-07T08:15:30.117 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:15:31.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:31 vm09 bash[21220]: cluster 2026-03-07T07:15:29.664535+0000 mgr.vm07.yrfcuj (mgr.14201) 1149 : cluster [DBG] pgmap v673: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:31.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:31 vm09 bash[21220]: cluster 2026-03-07T07:15:29.664535+0000 mgr.vm07.yrfcuj (mgr.14201) 1149 : cluster [DBG] pgmap v673: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:31.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:31 vm09 bash[21220]: audit 2026-03-07T07:15:29.749134+0000 mgr.vm07.yrfcuj (mgr.14201) 1150 : audit [DBG] from='client.16982 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:31.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:31 vm09 bash[21220]: audit 2026-03-07T07:15:29.749134+0000 mgr.vm07.yrfcuj (mgr.14201) 1150 : audit [DBG] from='client.16982 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:31.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:31 vm09 bash[21220]: audit 2026-03-07T07:15:29.924251+0000 mgr.vm07.yrfcuj (mgr.14201) 1151 : audit [DBG] from='client.16986 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:31.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:31 vm09 bash[21220]: audit 2026-03-07T07:15:29.924251+0000 mgr.vm07.yrfcuj (mgr.14201) 1151 : audit [DBG] from='client.16986 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:31.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:31 vm09 bash[21220]: audit 2026-03-07T07:15:30.123678+0000 mon.vm07 (mon.0) 1309 : audit [DBG] from='client.? 192.168.123.107:0/1557597196' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:31.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:31 vm09 bash[21220]: audit 2026-03-07T07:15:30.123678+0000 mon.vm07 (mon.0) 1309 : audit [DBG] from='client.? 192.168.123.107:0/1557597196' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:31.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:31 vm07 bash[17031]: cluster 2026-03-07T07:15:29.664535+0000 mgr.vm07.yrfcuj (mgr.14201) 1149 : cluster [DBG] pgmap v673: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:31.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:31 vm07 bash[17031]: cluster 2026-03-07T07:15:29.664535+0000 mgr.vm07.yrfcuj (mgr.14201) 1149 : cluster [DBG] pgmap v673: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:31.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:31 vm07 bash[17031]: audit 2026-03-07T07:15:29.749134+0000 mgr.vm07.yrfcuj (mgr.14201) 1150 : audit [DBG] from='client.16982 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:31.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:31 vm07 bash[17031]: audit 2026-03-07T07:15:29.749134+0000 mgr.vm07.yrfcuj (mgr.14201) 1150 : audit [DBG] from='client.16982 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:31.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:31 vm07 bash[17031]: audit 2026-03-07T07:15:29.924251+0000 mgr.vm07.yrfcuj (mgr.14201) 1151 : audit [DBG] from='client.16986 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:31.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:31 vm07 bash[17031]: audit 2026-03-07T07:15:29.924251+0000 mgr.vm07.yrfcuj (mgr.14201) 1151 : audit [DBG] from='client.16986 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:31.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:31 vm07 bash[17031]: audit 2026-03-07T07:15:30.123678+0000 mon.vm07 (mon.0) 1309 : audit [DBG] from='client.? 192.168.123.107:0/1557597196' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:31.397 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:31 vm07 bash[17031]: audit 2026-03-07T07:15:30.123678+0000 mon.vm07 (mon.0) 1309 : audit [DBG] from='client.? 192.168.123.107:0/1557597196' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:33.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:33 vm09 bash[21220]: cluster 2026-03-07T07:15:31.664948+0000 mgr.vm07.yrfcuj (mgr.14201) 1152 : cluster [DBG] pgmap v674: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:33.365 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:33 vm09 bash[21220]: cluster 2026-03-07T07:15:31.664948+0000 mgr.vm07.yrfcuj (mgr.14201) 1152 : cluster [DBG] pgmap v674: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:33.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:33 vm07 bash[17031]: cluster 2026-03-07T07:15:31.664948+0000 mgr.vm07.yrfcuj (mgr.14201) 1152 : cluster [DBG] pgmap v674: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:33.396 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:33 vm07 bash[17031]: cluster 2026-03-07T07:15:31.664948+0000 mgr.vm07.yrfcuj (mgr.14201) 1152 : cluster [DBG] pgmap v674: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:34 vm09 bash[21220]: audit 2026-03-07T07:15:33.291202+0000 mon.vm07 (mon.0) 1310 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:34 vm09 bash[21220]: audit 2026-03-07T07:15:33.291202+0000 mon.vm07 (mon.0) 1310 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:34 vm09 bash[21220]: audit 2026-03-07T07:15:33.297206+0000 mon.vm07 (mon.0) 1311 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:34 vm09 bash[21220]: audit 2026-03-07T07:15:33.297206+0000 mon.vm07 (mon.0) 1311 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:34 vm09 bash[21220]: audit 2026-03-07T07:15:33.298272+0000 mon.vm07 (mon.0) 1312 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:34 vm09 bash[21220]: audit 2026-03-07T07:15:33.298272+0000 mon.vm07 (mon.0) 1312 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:34 vm09 bash[21220]: audit 2026-03-07T07:15:33.299004+0000 mon.vm07 (mon.0) 1313 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:34 vm09 bash[21220]: audit 2026-03-07T07:15:33.299004+0000 mon.vm07 (mon.0) 1313 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:34 vm09 bash[21220]: audit 2026-03-07T07:15:33.303394+0000 mon.vm07 (mon.0) 1314 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:34 vm09 bash[21220]: audit 2026-03-07T07:15:33.303394+0000 mon.vm07 (mon.0) 1314 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:34 vm09 bash[21220]: audit 2026-03-07T07:15:33.304940+0000 mon.vm07 (mon.0) 1315 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:34.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:34 vm09 bash[21220]: audit 2026-03-07T07:15:33.304940+0000 mon.vm07 (mon.0) 1315 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:34.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:34 vm07 bash[17031]: audit 2026-03-07T07:15:33.291202+0000 mon.vm07 (mon.0) 1310 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:34.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:34 vm07 bash[17031]: audit 2026-03-07T07:15:33.291202+0000 mon.vm07 (mon.0) 1310 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:34.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:34 vm07 bash[17031]: audit 2026-03-07T07:15:33.297206+0000 mon.vm07 (mon.0) 1311 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:34.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:34 vm07 bash[17031]: audit 2026-03-07T07:15:33.297206+0000 mon.vm07 (mon.0) 1311 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:34.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:34 vm07 bash[17031]: audit 2026-03-07T07:15:33.298272+0000 mon.vm07 (mon.0) 1312 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:34.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:34 vm07 bash[17031]: audit 2026-03-07T07:15:33.298272+0000 mon.vm07 (mon.0) 1312 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:34.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:34 vm07 bash[17031]: audit 2026-03-07T07:15:33.299004+0000 mon.vm07 (mon.0) 1313 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:34.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:34 vm07 bash[17031]: audit 2026-03-07T07:15:33.299004+0000 mon.vm07 (mon.0) 1313 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:34.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:34 vm07 bash[17031]: audit 2026-03-07T07:15:33.303394+0000 mon.vm07 (mon.0) 1314 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:34.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:34 vm07 bash[17031]: audit 2026-03-07T07:15:33.303394+0000 mon.vm07 (mon.0) 1314 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:34.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:34 vm07 bash[17031]: audit 2026-03-07T07:15:33.304940+0000 mon.vm07 (mon.0) 1315 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:34.647 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:34 vm07 bash[17031]: audit 2026-03-07T07:15:33.304940+0000 mon.vm07 (mon.0) 1315 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:35.296 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm09.hhbnqk vm09 *:9000,9001 stopped 2s ago 17m - - 2026-03-07T08:15:35.306 INFO:teuthology.orchestra.run.vm07.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-07T08:15:35.308 INFO:teuthology.orchestra.run.vm07.stderr: Dload Upload Total Spent Left Speed 2026-03-07T08:15:35.312 INFO:teuthology.orchestra.run.vm07.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 37400 0 --:--:-- --:--:-- --:--:-- 37400 2026-03-07T08:15:35.485 INFO:teuthology.orchestra.run.vm07.stdout:anonymousScheduled to start haproxy.rgw.foo.vm09.hhbnqk on host 'vm09' 2026-03-07T08:15:35.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:35 vm09 bash[21220]: cluster 2026-03-07T07:15:33.665365+0000 mgr.vm07.yrfcuj (mgr.14201) 1153 : cluster [DBG] pgmap v675: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:15:35.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:35 vm09 bash[21220]: cluster 2026-03-07T07:15:33.665365+0000 mgr.vm07.yrfcuj (mgr.14201) 1153 : cluster [DBG] pgmap v675: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:15:35.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:35 vm07 bash[17031]: cluster 2026-03-07T07:15:33.665365+0000 mgr.vm07.yrfcuj (mgr.14201) 1153 : cluster [DBG] pgmap v675: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:15:35.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:35 vm07 bash[17031]: cluster 2026-03-07T07:15:33.665365+0000 mgr.vm07.yrfcuj (mgr.14201) 1153 : cluster [DBG] pgmap v675: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:15:35.670 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for haproxy.rgw.foo.vm09.hhbnqk to start 2026-03-07T08:15:35.816 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:15:35.816 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm07.rzyepz vm07 *:9000,9001 running (19s) 15s ago 17m 3516k - 2.3.17-d1c9119 e85424b0d443 5def5ff2697f 2026-03-07T08:15:35.816 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm09.hhbnqk vm09 *:9000,9001 stopped 2s ago 17m - - 2026-03-07T08:15:36.001 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:35.286205+0000 mgr.vm07.yrfcuj (mgr.14201) 1154 : audit [DBG] from='client.16994 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:35.286205+0000 mgr.vm07.yrfcuj (mgr.14201) 1154 : audit [DBG] from='client.16994 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:35.477064+0000 mgr.vm07.yrfcuj (mgr.14201) 1155 : audit [DBG] from='client.16998 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "haproxy.rgw.foo.vm09.hhbnqk", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:35.477064+0000 mgr.vm07.yrfcuj (mgr.14201) 1155 : audit [DBG] from='client.16998 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "haproxy.rgw.foo.vm09.hhbnqk", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: cephadm 2026-03-07T07:15:35.477414+0000 mgr.vm07.yrfcuj (mgr.14201) 1156 : cephadm [INF] Schedule start daemon haproxy.rgw.foo.vm09.hhbnqk 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: cephadm 2026-03-07T07:15:35.477414+0000 mgr.vm07.yrfcuj (mgr.14201) 1156 : cephadm [INF] Schedule start daemon haproxy.rgw.foo.vm09.hhbnqk 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:35.483534+0000 mon.vm07 (mon.0) 1316 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:35.483534+0000 mon.vm07 (mon.0) 1316 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:35.489334+0000 mon.vm07 (mon.0) 1317 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:35.489334+0000 mon.vm07 (mon.0) 1317 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:35.490483+0000 mon.vm07 (mon.0) 1318 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:35.490483+0000 mon.vm07 (mon.0) 1318 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:35.492057+0000 mon.vm07 (mon.0) 1319 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:35.492057+0000 mon.vm07 (mon.0) 1319 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:35.492746+0000 mon.vm07 (mon.0) 1320 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:35.492746+0000 mon.vm07 (mon.0) 1320 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:35.497375+0000 mon.vm07 (mon.0) 1321 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:35.497375+0000 mon.vm07 (mon.0) 1321 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:35.499055+0000 mon.vm07 (mon.0) 1322 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:35.499055+0000 mon.vm07 (mon.0) 1322 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:36.007755+0000 mon.vm07 (mon.0) 1323 : audit [DBG] from='client.? 192.168.123.107:0/1338724045' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:36.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:36 vm09 bash[21220]: audit 2026-03-07T07:15:36.007755+0000 mon.vm07 (mon.0) 1323 : audit [DBG] from='client.? 192.168.123.107:0/1338724045' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:36.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:35.286205+0000 mgr.vm07.yrfcuj (mgr.14201) 1154 : audit [DBG] from='client.16994 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:36.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:35.286205+0000 mgr.vm07.yrfcuj (mgr.14201) 1154 : audit [DBG] from='client.16994 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:36.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:35.477064+0000 mgr.vm07.yrfcuj (mgr.14201) 1155 : audit [DBG] from='client.16998 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "haproxy.rgw.foo.vm09.hhbnqk", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:36.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:35.477064+0000 mgr.vm07.yrfcuj (mgr.14201) 1155 : audit [DBG] from='client.16998 -' entity='client.admin' cmd=[{"prefix": "orch daemon", "action": "start", "name": "haproxy.rgw.foo.vm09.hhbnqk", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:36.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: cephadm 2026-03-07T07:15:35.477414+0000 mgr.vm07.yrfcuj (mgr.14201) 1156 : cephadm [INF] Schedule start daemon haproxy.rgw.foo.vm09.hhbnqk 2026-03-07T08:15:36.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: cephadm 2026-03-07T07:15:35.477414+0000 mgr.vm07.yrfcuj (mgr.14201) 1156 : cephadm [INF] Schedule start daemon haproxy.rgw.foo.vm09.hhbnqk 2026-03-07T08:15:36.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:35.483534+0000 mon.vm07 (mon.0) 1316 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:36.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:35.483534+0000 mon.vm07 (mon.0) 1316 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:36.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:35.489334+0000 mon.vm07 (mon.0) 1317 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:36.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:35.489334+0000 mon.vm07 (mon.0) 1317 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:36.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:35.490483+0000 mon.vm07 (mon.0) 1318 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:36.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:35.490483+0000 mon.vm07 (mon.0) 1318 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:36.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:35.492057+0000 mon.vm07 (mon.0) 1319 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:36.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:35.492057+0000 mon.vm07 (mon.0) 1319 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:36.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:35.492746+0000 mon.vm07 (mon.0) 1320 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:36.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:35.492746+0000 mon.vm07 (mon.0) 1320 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:36.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:35.497375+0000 mon.vm07 (mon.0) 1321 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:36.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:35.497375+0000 mon.vm07 (mon.0) 1321 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:36.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:35.499055+0000 mon.vm07 (mon.0) 1322 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:36.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:35.499055+0000 mon.vm07 (mon.0) 1322 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:36.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:36.007755+0000 mon.vm07 (mon.0) 1323 : audit [DBG] from='client.? 192.168.123.107:0/1338724045' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:36.897 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:36 vm07 bash[17031]: audit 2026-03-07T07:15:36.007755+0000 mon.vm07 (mon.0) 1323 : audit [DBG] from='client.? 192.168.123.107:0/1338724045' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:37 vm09 bash[21220]: audit 2026-03-07T07:15:35.660964+0000 mgr.vm07.yrfcuj (mgr.14201) 1157 : audit [DBG] from='client.17002 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:37 vm09 bash[21220]: audit 2026-03-07T07:15:35.660964+0000 mgr.vm07.yrfcuj (mgr.14201) 1157 : audit [DBG] from='client.17002 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:37 vm09 bash[21220]: cluster 2026-03-07T07:15:35.665677+0000 mgr.vm07.yrfcuj (mgr.14201) 1158 : cluster [DBG] pgmap v676: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:37 vm09 bash[21220]: cluster 2026-03-07T07:15:35.665677+0000 mgr.vm07.yrfcuj (mgr.14201) 1158 : cluster [DBG] pgmap v676: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:37 vm09 bash[21220]: audit 2026-03-07T07:15:35.820343+0000 mgr.vm07.yrfcuj (mgr.14201) 1159 : audit [DBG] from='client.17006 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:37.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:37 vm09 bash[21220]: audit 2026-03-07T07:15:35.820343+0000 mgr.vm07.yrfcuj (mgr.14201) 1159 : audit [DBG] from='client.17006 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:37.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:37 vm07 bash[17031]: audit 2026-03-07T07:15:35.660964+0000 mgr.vm07.yrfcuj (mgr.14201) 1157 : audit [DBG] from='client.17002 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:37.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:37 vm07 bash[17031]: audit 2026-03-07T07:15:35.660964+0000 mgr.vm07.yrfcuj (mgr.14201) 1157 : audit [DBG] from='client.17002 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:37.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:37 vm07 bash[17031]: cluster 2026-03-07T07:15:35.665677+0000 mgr.vm07.yrfcuj (mgr.14201) 1158 : cluster [DBG] pgmap v676: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:37.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:37 vm07 bash[17031]: cluster 2026-03-07T07:15:35.665677+0000 mgr.vm07.yrfcuj (mgr.14201) 1158 : cluster [DBG] pgmap v676: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:37.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:37 vm07 bash[17031]: audit 2026-03-07T07:15:35.820343+0000 mgr.vm07.yrfcuj (mgr.14201) 1159 : audit [DBG] from='client.17006 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:37.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:37 vm07 bash[17031]: audit 2026-03-07T07:15:35.820343+0000 mgr.vm07.yrfcuj (mgr.14201) 1159 : audit [DBG] from='client.17006 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:39.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:39 vm07 bash[17031]: cluster 2026-03-07T07:15:37.666139+0000 mgr.vm07.yrfcuj (mgr.14201) 1160 : cluster [DBG] pgmap v677: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:39.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:39 vm07 bash[17031]: cluster 2026-03-07T07:15:37.666139+0000 mgr.vm07.yrfcuj (mgr.14201) 1160 : cluster [DBG] pgmap v677: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:39 vm09 bash[21220]: cluster 2026-03-07T07:15:37.666139+0000 mgr.vm07.yrfcuj (mgr.14201) 1160 : cluster [DBG] pgmap v677: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:40.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:39 vm09 bash[21220]: cluster 2026-03-07T07:15:37.666139+0000 mgr.vm07.yrfcuj (mgr.14201) 1160 : cluster [DBG] pgmap v677: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:40.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:40 vm07 bash[17031]: audit 2026-03-07T07:15:40.148555+0000 mon.vm07 (mon.0) 1324 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:15:40.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:40 vm07 bash[17031]: audit 2026-03-07T07:15:40.148555+0000 mon.vm07 (mon.0) 1324 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:15:41.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:40 vm09 bash[21220]: audit 2026-03-07T07:15:40.148555+0000 mon.vm07 (mon.0) 1324 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:15:41.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:40 vm09 bash[21220]: audit 2026-03-07T07:15:40.148555+0000 mon.vm07 (mon.0) 1324 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:15:41.192 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for haproxy.rgw.foo.vm09.hhbnqk to start 2026-03-07T08:15:41.352 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:15:41.352 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm07.rzyepz vm07 *:9000,9001 running (25s) 20s ago 17m 3516k - 2.3.17-d1c9119 e85424b0d443 5def5ff2697f 2026-03-07T08:15:41.352 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm09.hhbnqk vm09 *:9000,9001 stopped 8s ago 17m - - 2026-03-07T08:15:41.539 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:15:41.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:41 vm07 bash[17031]: cluster 2026-03-07T07:15:39.666459+0000 mgr.vm07.yrfcuj (mgr.14201) 1161 : cluster [DBG] pgmap v678: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:41.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:41 vm07 bash[17031]: cluster 2026-03-07T07:15:39.666459+0000 mgr.vm07.yrfcuj (mgr.14201) 1161 : cluster [DBG] pgmap v678: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:41.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:41 vm07 bash[17031]: audit 2026-03-07T07:15:41.545441+0000 mon.vm07 (mon.0) 1325 : audit [DBG] from='client.? 192.168.123.107:0/4204775284' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:41.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:41 vm07 bash[17031]: audit 2026-03-07T07:15:41.545441+0000 mon.vm07 (mon.0) 1325 : audit [DBG] from='client.? 192.168.123.107:0/4204775284' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:41 vm09 bash[21220]: cluster 2026-03-07T07:15:39.666459+0000 mgr.vm07.yrfcuj (mgr.14201) 1161 : cluster [DBG] pgmap v678: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:41 vm09 bash[21220]: cluster 2026-03-07T07:15:39.666459+0000 mgr.vm07.yrfcuj (mgr.14201) 1161 : cluster [DBG] pgmap v678: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:41 vm09 bash[21220]: audit 2026-03-07T07:15:41.545441+0000 mon.vm07 (mon.0) 1325 : audit [DBG] from='client.? 192.168.123.107:0/4204775284' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:42.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:41 vm09 bash[21220]: audit 2026-03-07T07:15:41.545441+0000 mon.vm07 (mon.0) 1325 : audit [DBG] from='client.? 192.168.123.107:0/4204775284' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:42.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:42 vm07 bash[17031]: audit 2026-03-07T07:15:41.179680+0000 mgr.vm07.yrfcuj (mgr.14201) 1162 : audit [DBG] from='client.17014 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:42.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:42 vm07 bash[17031]: audit 2026-03-07T07:15:41.179680+0000 mgr.vm07.yrfcuj (mgr.14201) 1162 : audit [DBG] from='client.17014 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:42.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:42 vm07 bash[17031]: audit 2026-03-07T07:15:41.356189+0000 mgr.vm07.yrfcuj (mgr.14201) 1163 : audit [DBG] from='client.17018 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:42.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:42 vm07 bash[17031]: audit 2026-03-07T07:15:41.356189+0000 mgr.vm07.yrfcuj (mgr.14201) 1163 : audit [DBG] from='client.17018 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:43.108 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:42 vm09 bash[21220]: audit 2026-03-07T07:15:41.179680+0000 mgr.vm07.yrfcuj (mgr.14201) 1162 : audit [DBG] from='client.17014 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:43.108 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:42 vm09 bash[21220]: audit 2026-03-07T07:15:41.179680+0000 mgr.vm07.yrfcuj (mgr.14201) 1162 : audit [DBG] from='client.17014 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:43.108 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:42 vm09 bash[21220]: audit 2026-03-07T07:15:41.356189+0000 mgr.vm07.yrfcuj (mgr.14201) 1163 : audit [DBG] from='client.17018 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:43.108 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:42 vm09 bash[21220]: audit 2026-03-07T07:15:41.356189+0000 mgr.vm07.yrfcuj (mgr.14201) 1163 : audit [DBG] from='client.17018 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:44.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:43 vm09 bash[21220]: cluster 2026-03-07T07:15:41.666832+0000 mgr.vm07.yrfcuj (mgr.14201) 1164 : cluster [DBG] pgmap v679: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:44.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:43 vm09 bash[21220]: cluster 2026-03-07T07:15:41.666832+0000 mgr.vm07.yrfcuj (mgr.14201) 1164 : cluster [DBG] pgmap v679: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:44.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:43 vm07 bash[17031]: cluster 2026-03-07T07:15:41.666832+0000 mgr.vm07.yrfcuj (mgr.14201) 1164 : cluster [DBG] pgmap v679: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:44.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:43 vm07 bash[17031]: cluster 2026-03-07T07:15:41.666832+0000 mgr.vm07.yrfcuj (mgr.14201) 1164 : cluster [DBG] pgmap v679: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:45 vm09 bash[21220]: cluster 2026-03-07T07:15:43.667293+0000 mgr.vm07.yrfcuj (mgr.14201) 1165 : cluster [DBG] pgmap v680: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:45 vm09 bash[21220]: cluster 2026-03-07T07:15:43.667293+0000 mgr.vm07.yrfcuj (mgr.14201) 1165 : cluster [DBG] pgmap v680: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:45 vm09 bash[21220]: audit 2026-03-07T07:15:44.356774+0000 mon.vm07 (mon.0) 1326 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:45 vm09 bash[21220]: audit 2026-03-07T07:15:44.356774+0000 mon.vm07 (mon.0) 1326 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:45 vm09 bash[21220]: audit 2026-03-07T07:15:44.363395+0000 mon.vm07 (mon.0) 1327 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:45 vm09 bash[21220]: audit 2026-03-07T07:15:44.363395+0000 mon.vm07 (mon.0) 1327 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:45 vm09 bash[21220]: audit 2026-03-07T07:15:44.391000+0000 mon.vm07 (mon.0) 1328 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:45.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:45 vm09 bash[21220]: audit 2026-03-07T07:15:44.391000+0000 mon.vm07 (mon.0) 1328 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:45.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:45 vm07 bash[17031]: cluster 2026-03-07T07:15:43.667293+0000 mgr.vm07.yrfcuj (mgr.14201) 1165 : cluster [DBG] pgmap v680: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:45.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:45 vm07 bash[17031]: cluster 2026-03-07T07:15:43.667293+0000 mgr.vm07.yrfcuj (mgr.14201) 1165 : cluster [DBG] pgmap v680: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:45.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:45 vm07 bash[17031]: audit 2026-03-07T07:15:44.356774+0000 mon.vm07 (mon.0) 1326 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:45.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:45 vm07 bash[17031]: audit 2026-03-07T07:15:44.356774+0000 mon.vm07 (mon.0) 1326 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:45.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:45 vm07 bash[17031]: audit 2026-03-07T07:15:44.363395+0000 mon.vm07 (mon.0) 1327 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:45.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:45 vm07 bash[17031]: audit 2026-03-07T07:15:44.363395+0000 mon.vm07 (mon.0) 1327 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:45.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:45 vm07 bash[17031]: audit 2026-03-07T07:15:44.391000+0000 mon.vm07 (mon.0) 1328 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:45.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:45 vm07 bash[17031]: audit 2026-03-07T07:15:44.391000+0000 mon.vm07 (mon.0) 1328 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-07T08:15:46.708 INFO:teuthology.orchestra.run.vm07.stdout:Waiting for haproxy.rgw.foo.vm09.hhbnqk to start 2026-03-07T08:15:46.858 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:15:46.858 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm07.rzyepz vm07 *:9000,9001 running (30s) 26s ago 17m 3516k - 2.3.17-d1c9119 e85424b0d443 5def5ff2697f 2026-03-07T08:15:46.858 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm09.hhbnqk vm09 *:9000,9001 stopped 13s ago 18m - - 2026-03-07T08:15:47.054 INFO:teuthology.orchestra.run.vm07.stdout:HEALTH_OK 2026-03-07T08:15:47.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:47 vm09 bash[21220]: cluster 2026-03-07T07:15:45.667616+0000 mgr.vm07.yrfcuj (mgr.14201) 1166 : cluster [DBG] pgmap v681: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:47.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:47 vm09 bash[21220]: cluster 2026-03-07T07:15:45.667616+0000 mgr.vm07.yrfcuj (mgr.14201) 1166 : cluster [DBG] pgmap v681: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:47.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:47 vm09 bash[21220]: audit 2026-03-07T07:15:47.060386+0000 mon.vm07 (mon.0) 1329 : audit [DBG] from='client.? 192.168.123.107:0/2814973772' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:47.615 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:47 vm09 bash[21220]: audit 2026-03-07T07:15:47.060386+0000 mon.vm07 (mon.0) 1329 : audit [DBG] from='client.? 192.168.123.107:0/2814973772' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:47.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:47 vm07 bash[17031]: cluster 2026-03-07T07:15:45.667616+0000 mgr.vm07.yrfcuj (mgr.14201) 1166 : cluster [DBG] pgmap v681: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:47.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:47 vm07 bash[17031]: cluster 2026-03-07T07:15:45.667616+0000 mgr.vm07.yrfcuj (mgr.14201) 1166 : cluster [DBG] pgmap v681: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:47.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:47 vm07 bash[17031]: audit 2026-03-07T07:15:47.060386+0000 mon.vm07 (mon.0) 1329 : audit [DBG] from='client.? 192.168.123.107:0/2814973772' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:47.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:47 vm07 bash[17031]: audit 2026-03-07T07:15:47.060386+0000 mon.vm07 (mon.0) 1329 : audit [DBG] from='client.? 192.168.123.107:0/2814973772' entity='client.admin' cmd=[{"prefix": "health", "detail": "detail"}]: dispatch 2026-03-07T08:15:48.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:48 vm07 bash[17031]: audit 2026-03-07T07:15:46.700664+0000 mgr.vm07.yrfcuj (mgr.14201) 1167 : audit [DBG] from='client.17026 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:48.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:48 vm07 bash[17031]: audit 2026-03-07T07:15:46.700664+0000 mgr.vm07.yrfcuj (mgr.14201) 1167 : audit [DBG] from='client.17026 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:48.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:48 vm07 bash[17031]: audit 2026-03-07T07:15:46.862435+0000 mgr.vm07.yrfcuj (mgr.14201) 1168 : audit [DBG] from='client.17030 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:48.646 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:48 vm07 bash[17031]: audit 2026-03-07T07:15:46.862435+0000 mgr.vm07.yrfcuj (mgr.14201) 1168 : audit [DBG] from='client.17030 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:48 vm09 bash[21220]: audit 2026-03-07T07:15:46.700664+0000 mgr.vm07.yrfcuj (mgr.14201) 1167 : audit [DBG] from='client.17026 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:48 vm09 bash[21220]: audit 2026-03-07T07:15:46.700664+0000 mgr.vm07.yrfcuj (mgr.14201) 1167 : audit [DBG] from='client.17026 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:48 vm09 bash[21220]: audit 2026-03-07T07:15:46.862435+0000 mgr.vm07.yrfcuj (mgr.14201) 1168 : audit [DBG] from='client.17030 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:48.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:48 vm09 bash[21220]: audit 2026-03-07T07:15:46.862435+0000 mgr.vm07.yrfcuj (mgr.14201) 1168 : audit [DBG] from='client.17030 -' entity='client.admin' cmd=[{"prefix": "orch ps", "daemon_type": "haproxy", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:49 vm09 bash[21220]: cluster 2026-03-07T07:15:47.668146+0000 mgr.vm07.yrfcuj (mgr.14201) 1169 : cluster [DBG] pgmap v682: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:49 vm09 bash[21220]: cluster 2026-03-07T07:15:47.668146+0000 mgr.vm07.yrfcuj (mgr.14201) 1169 : cluster [DBG] pgmap v682: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:49 vm09 bash[21220]: audit 2026-03-07T07:15:48.756230+0000 mon.vm07 (mon.0) 1330 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:49 vm09 bash[21220]: audit 2026-03-07T07:15:48.756230+0000 mon.vm07 (mon.0) 1330 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:49 vm09 bash[21220]: audit 2026-03-07T07:15:48.761112+0000 mon.vm07 (mon.0) 1331 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:49 vm09 bash[21220]: audit 2026-03-07T07:15:48.761112+0000 mon.vm07 (mon.0) 1331 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:49 vm09 bash[21220]: audit 2026-03-07T07:15:48.762017+0000 mon.vm07 (mon.0) 1332 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:49 vm09 bash[21220]: audit 2026-03-07T07:15:48.762017+0000 mon.vm07 (mon.0) 1332 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:49 vm09 bash[21220]: audit 2026-03-07T07:15:48.762486+0000 mon.vm07 (mon.0) 1333 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:49 vm09 bash[21220]: audit 2026-03-07T07:15:48.762486+0000 mon.vm07 (mon.0) 1333 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:49 vm09 bash[21220]: audit 2026-03-07T07:15:48.766345+0000 mon.vm07 (mon.0) 1334 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:49 vm09 bash[21220]: audit 2026-03-07T07:15:48.766345+0000 mon.vm07 (mon.0) 1334 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:49 vm09 bash[21220]: audit 2026-03-07T07:15:48.767694+0000 mon.vm07 (mon.0) 1335 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:49.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:49 vm09 bash[21220]: audit 2026-03-07T07:15:48.767694+0000 mon.vm07 (mon.0) 1335 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:49.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:49 vm07 bash[17031]: cluster 2026-03-07T07:15:47.668146+0000 mgr.vm07.yrfcuj (mgr.14201) 1169 : cluster [DBG] pgmap v682: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:49.916 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:49 vm07 bash[17031]: cluster 2026-03-07T07:15:47.668146+0000 mgr.vm07.yrfcuj (mgr.14201) 1169 : cluster [DBG] pgmap v682: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:49.916 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:49 vm07 bash[17031]: audit 2026-03-07T07:15:48.756230+0000 mon.vm07 (mon.0) 1330 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:49.916 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:49 vm07 bash[17031]: audit 2026-03-07T07:15:48.756230+0000 mon.vm07 (mon.0) 1330 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:49.916 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:49 vm07 bash[17031]: audit 2026-03-07T07:15:48.761112+0000 mon.vm07 (mon.0) 1331 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:49.916 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:49 vm07 bash[17031]: audit 2026-03-07T07:15:48.761112+0000 mon.vm07 (mon.0) 1331 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:49.916 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:49 vm07 bash[17031]: audit 2026-03-07T07:15:48.762017+0000 mon.vm07 (mon.0) 1332 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:49.916 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:49 vm07 bash[17031]: audit 2026-03-07T07:15:48.762017+0000 mon.vm07 (mon.0) 1332 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-07T08:15:49.916 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:49 vm07 bash[17031]: audit 2026-03-07T07:15:48.762486+0000 mon.vm07 (mon.0) 1333 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:49.916 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:49 vm07 bash[17031]: audit 2026-03-07T07:15:48.762486+0000 mon.vm07 (mon.0) 1333 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-07T08:15:49.916 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:49 vm07 bash[17031]: audit 2026-03-07T07:15:48.766345+0000 mon.vm07 (mon.0) 1334 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:49.916 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:49 vm07 bash[17031]: audit 2026-03-07T07:15:48.766345+0000 mon.vm07 (mon.0) 1334 : audit [INF] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' 2026-03-07T08:15:49.916 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:49 vm07 bash[17031]: audit 2026-03-07T07:15:48.767694+0000 mon.vm07 (mon.0) 1335 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:49.916 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:49 vm07 bash[17031]: audit 2026-03-07T07:15:48.767694+0000 mon.vm07 (mon.0) 1335 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-07T08:15:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:51 vm09 bash[21220]: cluster 2026-03-07T07:15:49.668510+0000 mgr.vm07.yrfcuj (mgr.14201) 1170 : cluster [DBG] pgmap v683: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:51.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:51 vm09 bash[21220]: cluster 2026-03-07T07:15:49.668510+0000 mgr.vm07.yrfcuj (mgr.14201) 1170 : cluster [DBG] pgmap v683: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:51.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:51 vm07 bash[17031]: cluster 2026-03-07T07:15:49.668510+0000 mgr.vm07.yrfcuj (mgr.14201) 1170 : cluster [DBG] pgmap v683: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:51.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:51 vm07 bash[17031]: cluster 2026-03-07T07:15:49.668510+0000 mgr.vm07.yrfcuj (mgr.14201) 1170 : cluster [DBG] pgmap v683: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:52.220 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm09.hhbnqk vm09 *:9000,9001 running (7s) 3s ago 18m 3512k - 2.3.17-d1c9119 e85424b0d443 a6b0a8925a4b 2026-03-07T08:15:52.225 INFO:teuthology.orchestra.run.vm07.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-03-07T08:15:52.225 INFO:teuthology.orchestra.run.vm07.stderr: Dload Upload Total Spent Left Speed 2026-03-07T08:15:52.225 INFO:teuthology.orchestra.run.vm07.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 182k 0 --:--:-- --:--:-- --:--:-- 182k 2026-03-07T08:15:52.281 INFO:teuthology.orchestra.run.vm07.stdout:anonymous 2026-03-07T08:15:52.281 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-07T08:15:52.283 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm07.local 2026-03-07T08:15:52.283 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- bash -c 'stat -c '"'"'%u %g'"'"' /var/log/ceph | grep '"'"'167 167'"'"'' 2026-03-07T08:15:53.864 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:53 vm09 bash[21220]: cluster 2026-03-07T07:15:51.668893+0000 mgr.vm07.yrfcuj (mgr.14201) 1171 : cluster [DBG] pgmap v684: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:53.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:53 vm09 bash[21220]: cluster 2026-03-07T07:15:51.668893+0000 mgr.vm07.yrfcuj (mgr.14201) 1171 : cluster [DBG] pgmap v684: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:53.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:53 vm09 bash[21220]: audit 2026-03-07T07:15:52.211963+0000 mgr.vm07.yrfcuj (mgr.14201) 1172 : audit [DBG] from='client.17038 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:53.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:53 vm09 bash[21220]: audit 2026-03-07T07:15:52.211963+0000 mgr.vm07.yrfcuj (mgr.14201) 1172 : audit [DBG] from='client.17038 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:53.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:53 vm07 bash[17031]: cluster 2026-03-07T07:15:51.668893+0000 mgr.vm07.yrfcuj (mgr.14201) 1171 : cluster [DBG] pgmap v684: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:53.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:53 vm07 bash[17031]: cluster 2026-03-07T07:15:51.668893+0000 mgr.vm07.yrfcuj (mgr.14201) 1171 : cluster [DBG] pgmap v684: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:53.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:53 vm07 bash[17031]: audit 2026-03-07T07:15:52.211963+0000 mgr.vm07.yrfcuj (mgr.14201) 1172 : audit [DBG] from='client.17038 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:53.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:53 vm07 bash[17031]: audit 2026-03-07T07:15:52.211963+0000 mgr.vm07.yrfcuj (mgr.14201) 1172 : audit [DBG] from='client.17038 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:15:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:55 vm09 bash[21220]: cluster 2026-03-07T07:15:53.669354+0000 mgr.vm07.yrfcuj (mgr.14201) 1173 : cluster [DBG] pgmap v685: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:55 vm09 bash[21220]: cluster 2026-03-07T07:15:53.669354+0000 mgr.vm07.yrfcuj (mgr.14201) 1173 : cluster [DBG] pgmap v685: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:55 vm09 bash[21220]: audit 2026-03-07T07:15:55.148683+0000 mon.vm07 (mon.0) 1336 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:15:55.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:55 vm09 bash[21220]: audit 2026-03-07T07:15:55.148683+0000 mon.vm07 (mon.0) 1336 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:15:55.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:55 vm07 bash[17031]: cluster 2026-03-07T07:15:53.669354+0000 mgr.vm07.yrfcuj (mgr.14201) 1173 : cluster [DBG] pgmap v685: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:55.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:55 vm07 bash[17031]: cluster 2026-03-07T07:15:53.669354+0000 mgr.vm07.yrfcuj (mgr.14201) 1173 : cluster [DBG] pgmap v685: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:55.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:55 vm07 bash[17031]: audit 2026-03-07T07:15:55.148683+0000 mon.vm07 (mon.0) 1336 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:15:55.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:55 vm07 bash[17031]: audit 2026-03-07T07:15:55.148683+0000 mon.vm07 (mon.0) 1336 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:15:57.060 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T08:15:57.157 INFO:teuthology.orchestra.run.vm07.stdout:167 167 2026-03-07T08:15:57.204 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- bash -c 'ceph orch status' 2026-03-07T08:15:57.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:57 vm09 bash[21220]: cluster 2026-03-07T07:15:55.669785+0000 mgr.vm07.yrfcuj (mgr.14201) 1174 : cluster [DBG] pgmap v686: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:57.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:57 vm09 bash[21220]: cluster 2026-03-07T07:15:55.669785+0000 mgr.vm07.yrfcuj (mgr.14201) 1174 : cluster [DBG] pgmap v686: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:57.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:57 vm07 bash[17031]: cluster 2026-03-07T07:15:55.669785+0000 mgr.vm07.yrfcuj (mgr.14201) 1174 : cluster [DBG] pgmap v686: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:57.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:57 vm07 bash[17031]: cluster 2026-03-07T07:15:55.669785+0000 mgr.vm07.yrfcuj (mgr.14201) 1174 : cluster [DBG] pgmap v686: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:15:59.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:59 vm09 bash[21220]: cluster 2026-03-07T07:15:57.670166+0000 mgr.vm07.yrfcuj (mgr.14201) 1175 : cluster [DBG] pgmap v687: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:59.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:15:59 vm09 bash[21220]: cluster 2026-03-07T07:15:57.670166+0000 mgr.vm07.yrfcuj (mgr.14201) 1175 : cluster [DBG] pgmap v687: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:59.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:59 vm07 bash[17031]: cluster 2026-03-07T07:15:57.670166+0000 mgr.vm07.yrfcuj (mgr.14201) 1175 : cluster [DBG] pgmap v687: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:15:59.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:15:59 vm07 bash[17031]: cluster 2026-03-07T07:15:57.670166+0000 mgr.vm07.yrfcuj (mgr.14201) 1175 : cluster [DBG] pgmap v687: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:01.098 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T08:16:01.447 INFO:teuthology.orchestra.run.vm07.stdout:Backend: cephadm 2026-03-07T08:16:01.447 INFO:teuthology.orchestra.run.vm07.stdout:Available: Yes 2026-03-07T08:16:01.447 INFO:teuthology.orchestra.run.vm07.stdout:Paused: No 2026-03-07T08:16:01.513 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- bash -c 'ceph orch ps' 2026-03-07T08:16:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:01 vm09 bash[21220]: cluster 2026-03-07T07:15:59.670495+0000 mgr.vm07.yrfcuj (mgr.14201) 1176 : cluster [DBG] pgmap v688: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:01.865 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:01 vm09 bash[21220]: cluster 2026-03-07T07:15:59.670495+0000 mgr.vm07.yrfcuj (mgr.14201) 1176 : cluster [DBG] pgmap v688: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:01.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:01 vm07 bash[17031]: cluster 2026-03-07T07:15:59.670495+0000 mgr.vm07.yrfcuj (mgr.14201) 1176 : cluster [DBG] pgmap v688: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:01.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:01 vm07 bash[17031]: cluster 2026-03-07T07:15:59.670495+0000 mgr.vm07.yrfcuj (mgr.14201) 1176 : cluster [DBG] pgmap v688: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:02.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:02 vm07 bash[17031]: audit 2026-03-07T07:16:01.454004+0000 mgr.vm07.yrfcuj (mgr.14201) 1177 : audit [DBG] from='client.17042 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:02.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:02 vm07 bash[17031]: audit 2026-03-07T07:16:01.454004+0000 mgr.vm07.yrfcuj (mgr.14201) 1177 : audit [DBG] from='client.17042 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:03.108 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:02 vm09 bash[21220]: audit 2026-03-07T07:16:01.454004+0000 mgr.vm07.yrfcuj (mgr.14201) 1177 : audit [DBG] from='client.17042 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:03.108 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:02 vm09 bash[21220]: audit 2026-03-07T07:16:01.454004+0000 mgr.vm07.yrfcuj (mgr.14201) 1177 : audit [DBG] from='client.17042 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:03.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:03 vm07 bash[17031]: cluster 2026-03-07T07:16:01.670842+0000 mgr.vm07.yrfcuj (mgr.14201) 1178 : cluster [DBG] pgmap v689: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:03.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:03 vm07 bash[17031]: cluster 2026-03-07T07:16:01.670842+0000 mgr.vm07.yrfcuj (mgr.14201) 1178 : cluster [DBG] pgmap v689: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:04.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:03 vm09 bash[21220]: cluster 2026-03-07T07:16:01.670842+0000 mgr.vm07.yrfcuj (mgr.14201) 1178 : cluster [DBG] pgmap v689: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:04.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:03 vm09 bash[21220]: cluster 2026-03-07T07:16:01.670842+0000 mgr.vm07.yrfcuj (mgr.14201) 1178 : cluster [DBG] pgmap v689: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:05.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:05 vm07 bash[17031]: cluster 2026-03-07T07:16:03.671306+0000 mgr.vm07.yrfcuj (mgr.14201) 1179 : cluster [DBG] pgmap v690: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:16:05.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:05 vm07 bash[17031]: cluster 2026-03-07T07:16:03.671306+0000 mgr.vm07.yrfcuj (mgr.14201) 1179 : cluster [DBG] pgmap v690: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:16:06.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:05 vm09 bash[21220]: cluster 2026-03-07T07:16:03.671306+0000 mgr.vm07.yrfcuj (mgr.14201) 1179 : cluster [DBG] pgmap v690: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:16:06.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:05 vm09 bash[21220]: cluster 2026-03-07T07:16:03.671306+0000 mgr.vm07.yrfcuj (mgr.14201) 1179 : cluster [DBG] pgmap v690: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 255 B/s rd, 511 B/s wr, 0 op/s 2026-03-07T08:16:06.299 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T08:16:06.647 INFO:teuthology.orchestra.run.vm07.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-07T08:16:06.647 INFO:teuthology.orchestra.run.vm07.stdout:alertmanager.vm07 vm07 *:9093,9094 running (21m) 45s ago 22m 13.6M - 0.25.0 c8568f914cd2 ced8afb39df2 2026-03-07T08:16:06.647 INFO:teuthology.orchestra.run.vm07.stdout:ceph-exporter.vm07 vm07 running (22m) 45s ago 22m 10.4M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 65356e296978 2026-03-07T08:16:06.647 INFO:teuthology.orchestra.run.vm07.stdout:ceph-exporter.vm09 vm09 running (21m) 17s ago 21m 8047k - 19.2.3-39-g340d3c24fc6 8bccc98d839a ef35692c433d 2026-03-07T08:16:06.647 INFO:teuthology.orchestra.run.vm07.stdout:crash.vm07 vm07 running (22m) 45s ago 22m 10.7M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c38c18a5b965 2026-03-07T08:16:06.647 INFO:teuthology.orchestra.run.vm07.stdout:crash.vm09 vm09 running (21m) 17s ago 21m 10.7M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 16d01acf9cbb 2026-03-07T08:16:06.647 INFO:teuthology.orchestra.run.vm07.stdout:grafana.vm07 vm07 *:3000 running (21m) 45s ago 21m 64.5M - 10.4.0 c8b91775d855 fbed394a229c 2026-03-07T08:16:06.647 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm07.rzyepz vm07 *:9000,9001 running (50s) 45s ago 18m 3516k - 2.3.17-d1c9119 e85424b0d443 5def5ff2697f 2026-03-07T08:16:06.647 INFO:teuthology.orchestra.run.vm07.stdout:haproxy.rgw.foo.vm09.hhbnqk vm09 *:9000,9001 running (22s) 17s ago 18m 3512k - 2.3.17-d1c9119 e85424b0d443 a6b0a8925a4b 2026-03-07T08:16:06.647 INFO:teuthology.orchestra.run.vm07.stdout:keepalived.rgw.foo.vm07.mbvpkm vm07 running (18m) 45s ago 18m 2503k - 2.2.4 4a3a1ff181d9 e8520e332bb7 2026-03-07T08:16:06.647 INFO:teuthology.orchestra.run.vm07.stdout:keepalived.rgw.foo.vm09.bnmvqb vm09 running (18m) 17s ago 18m 2476k - 2.2.4 4a3a1ff181d9 0654200d01b0 2026-03-07T08:16:06.647 INFO:teuthology.orchestra.run.vm07.stdout:mgr.vm07.yrfcuj vm07 *:9283,8765,8443 running (23m) 45s ago 23m 546M - 19.2.3-39-g340d3c24fc6 8bccc98d839a afb3597f2ccd 2026-03-07T08:16:06.647 INFO:teuthology.orchestra.run.vm07.stdout:mgr.vm09.eqznpw vm09 *:8443,9283,8765 running (21m) 17s ago 21m 478M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 046b6d4953d0 2026-03-07T08:16:06.647 INFO:teuthology.orchestra.run.vm07.stdout:mon.vm07 vm07 running (23m) 45s ago 23m 69.5M 2048M 19.2.3-39-g340d3c24fc6 8bccc98d839a df7f9a962253 2026-03-07T08:16:06.647 INFO:teuthology.orchestra.run.vm07.stdout:mon.vm09 vm09 running (21m) 17s ago 21m 43.5M 2048M 19.2.3-39-g340d3c24fc6 8bccc98d839a ef7ec53aa0b8 2026-03-07T08:16:06.647 INFO:teuthology.orchestra.run.vm07.stdout:node-exporter.vm07 vm07 *:9100 running (22m) 45s ago 22m 8124k - 1.7.0 72c9c2088986 1859f95972e0 2026-03-07T08:16:06.647 INFO:teuthology.orchestra.run.vm07.stdout:node-exporter.vm09 vm09 *:9100 running (21m) 17s ago 21m 8064k - 1.7.0 72c9c2088986 7e9cc4e2059c 2026-03-07T08:16:06.647 INFO:teuthology.orchestra.run.vm07.stdout:osd.0 vm09 running (20m) 17s ago 20m 74.5M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 2b2c1a047d82 2026-03-07T08:16:06.647 INFO:teuthology.orchestra.run.vm07.stdout:osd.1 vm07 running (20m) 45s ago 20m 56.9M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 5c199f0acc15 2026-03-07T08:16:06.648 INFO:teuthology.orchestra.run.vm07.stdout:osd.2 vm07 running (20m) 45s ago 20m 70.6M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a ff80e2683136 2026-03-07T08:16:06.648 INFO:teuthology.orchestra.run.vm07.stdout:osd.3 vm09 running (20m) 17s ago 20m 78.8M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 75b3f3ac04da 2026-03-07T08:16:06.648 INFO:teuthology.orchestra.run.vm07.stdout:osd.4 vm09 running (20m) 17s ago 20m 75.2M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 527a66f352ad 2026-03-07T08:16:06.648 INFO:teuthology.orchestra.run.vm07.stdout:osd.5 vm07 running (20m) 45s ago 20m 73.9M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 5f3f015dd990 2026-03-07T08:16:06.648 INFO:teuthology.orchestra.run.vm07.stdout:osd.6 vm07 running (20m) 45s ago 20m 53.4M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a ce94a3b6ab7b 2026-03-07T08:16:06.648 INFO:teuthology.orchestra.run.vm07.stdout:osd.7 vm09 running (20m) 17s ago 20m 52.7M 4096M 19.2.3-39-g340d3c24fc6 8bccc98d839a 80f17c672a66 2026-03-07T08:16:06.648 INFO:teuthology.orchestra.run.vm07.stdout:prometheus.vm07 vm07 *:9095 running (17m) 45s ago 21m 54.5M - 2.51.0 1d3b7f56885b 9b5861ffde54 2026-03-07T08:16:06.648 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.bgfaka vm07 *:8000 running (12m) 45s ago 18m 125M - 19.2.3-39-g340d3c24fc6 8bccc98d839a bbd6754b26b8 2026-03-07T08:16:06.648 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm07.lrxyjy vm07 *:8001 running (11m) 45s ago 18m 125M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 439e2e8e5bb1 2026-03-07T08:16:06.648 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.fgzozy vm09 *:8000 running (6m) 17s ago 18m 105M - 19.2.3-39-g340d3c24fc6 8bccc98d839a c9d4a9e850eb 2026-03-07T08:16:06.648 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo.vm09.kpgoql vm09 *:8001 running (79s) 17s ago 18m 91.5M - 19.2.3-39-g340d3c24fc6 8bccc98d839a 8ac366e53c02 2026-03-07T08:16:06.709 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- bash -c 'ceph orch ls' 2026-03-07T08:16:07.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:07 vm07 bash[17031]: cluster 2026-03-07T07:16:05.671670+0000 mgr.vm07.yrfcuj (mgr.14201) 1180 : cluster [DBG] pgmap v691: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:07.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:07 vm07 bash[17031]: cluster 2026-03-07T07:16:05.671670+0000 mgr.vm07.yrfcuj (mgr.14201) 1180 : cluster [DBG] pgmap v691: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:07.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:07 vm07 bash[17031]: audit 2026-03-07T07:16:06.647320+0000 mgr.vm07.yrfcuj (mgr.14201) 1181 : audit [DBG] from='client.25965 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:07.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:07 vm07 bash[17031]: audit 2026-03-07T07:16:06.647320+0000 mgr.vm07.yrfcuj (mgr.14201) 1181 : audit [DBG] from='client.25965 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:08.108 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:07 vm09 bash[21220]: cluster 2026-03-07T07:16:05.671670+0000 mgr.vm07.yrfcuj (mgr.14201) 1180 : cluster [DBG] pgmap v691: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:08.109 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:07 vm09 bash[21220]: cluster 2026-03-07T07:16:05.671670+0000 mgr.vm07.yrfcuj (mgr.14201) 1180 : cluster [DBG] pgmap v691: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:08.109 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:07 vm09 bash[21220]: audit 2026-03-07T07:16:06.647320+0000 mgr.vm07.yrfcuj (mgr.14201) 1181 : audit [DBG] from='client.25965 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:08.109 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:07 vm09 bash[21220]: audit 2026-03-07T07:16:06.647320+0000 mgr.vm07.yrfcuj (mgr.14201) 1181 : audit [DBG] from='client.25965 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:09.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:09 vm07 bash[17031]: cluster 2026-03-07T07:16:07.672087+0000 mgr.vm07.yrfcuj (mgr.14201) 1182 : cluster [DBG] pgmap v692: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:09.896 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:09 vm07 bash[17031]: cluster 2026-03-07T07:16:07.672087+0000 mgr.vm07.yrfcuj (mgr.14201) 1182 : cluster [DBG] pgmap v692: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:10.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:09 vm09 bash[21220]: cluster 2026-03-07T07:16:07.672087+0000 mgr.vm07.yrfcuj (mgr.14201) 1182 : cluster [DBG] pgmap v692: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:10.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:09 vm09 bash[21220]: cluster 2026-03-07T07:16:07.672087+0000 mgr.vm07.yrfcuj (mgr.14201) 1182 : cluster [DBG] pgmap v692: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:11.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:10 vm09 bash[21220]: cluster 2026-03-07T07:16:09.672403+0000 mgr.vm07.yrfcuj (mgr.14201) 1183 : cluster [DBG] pgmap v693: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:11.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:10 vm09 bash[21220]: cluster 2026-03-07T07:16:09.672403+0000 mgr.vm07.yrfcuj (mgr.14201) 1183 : cluster [DBG] pgmap v693: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:11.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:10 vm09 bash[21220]: audit 2026-03-07T07:16:10.149053+0000 mon.vm07 (mon.0) 1337 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:16:11.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:10 vm09 bash[21220]: audit 2026-03-07T07:16:10.149053+0000 mon.vm07 (mon.0) 1337 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:16:11.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:10 vm07 bash[17031]: cluster 2026-03-07T07:16:09.672403+0000 mgr.vm07.yrfcuj (mgr.14201) 1183 : cluster [DBG] pgmap v693: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:11.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:10 vm07 bash[17031]: cluster 2026-03-07T07:16:09.672403+0000 mgr.vm07.yrfcuj (mgr.14201) 1183 : cluster [DBG] pgmap v693: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:11.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:10 vm07 bash[17031]: audit 2026-03-07T07:16:10.149053+0000 mon.vm07 (mon.0) 1337 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:16:11.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:10 vm07 bash[17031]: audit 2026-03-07T07:16:10.149053+0000 mon.vm07 (mon.0) 1337 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:16:11.490 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T08:16:11.849 INFO:teuthology.orchestra.run.vm07.stdout:NAME PORTS RUNNING REFRESHED AGE PLACEMENT 2026-03-07T08:16:11.849 INFO:teuthology.orchestra.run.vm07.stdout:alertmanager ?:9093,9094 1/1 51s ago 22m count:1 2026-03-07T08:16:11.849 INFO:teuthology.orchestra.run.vm07.stdout:ceph-exporter 2/2 51s ago 22m * 2026-03-07T08:16:11.849 INFO:teuthology.orchestra.run.vm07.stdout:crash 2/2 51s ago 22m * 2026-03-07T08:16:11.849 INFO:teuthology.orchestra.run.vm07.stdout:grafana ?:3000 1/1 51s ago 22m count:1 2026-03-07T08:16:11.849 INFO:teuthology.orchestra.run.vm07.stdout:ingress.rgw.foo 12.12.1.107:9000,9001 4/4 51s ago 18m count:2 2026-03-07T08:16:11.849 INFO:teuthology.orchestra.run.vm07.stdout:mgr 2/2 51s ago 22m count:2 2026-03-07T08:16:11.849 INFO:teuthology.orchestra.run.vm07.stdout:mon 2/2 51s ago 21m vm07:192.168.123.107=vm07;vm09:192.168.123.109=vm09;count:2 2026-03-07T08:16:11.849 INFO:teuthology.orchestra.run.vm07.stdout:node-exporter ?:9100 2/2 51s ago 22m * 2026-03-07T08:16:11.849 INFO:teuthology.orchestra.run.vm07.stdout:osd.all-available-devices 8 51s ago 21m * 2026-03-07T08:16:11.849 INFO:teuthology.orchestra.run.vm07.stdout:prometheus ?:9095 1/1 51s ago 22m count:1 2026-03-07T08:16:11.849 INFO:teuthology.orchestra.run.vm07.stdout:rgw.foo ?:8000 4/4 51s ago 18m count:4;* 2026-03-07T08:16:11.911 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- bash -c 'ceph orch host ls' 2026-03-07T08:16:13.109 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:12 vm09 bash[21220]: cluster 2026-03-07T07:16:11.672822+0000 mgr.vm07.yrfcuj (mgr.14201) 1184 : cluster [DBG] pgmap v694: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:13.109 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:12 vm09 bash[21220]: cluster 2026-03-07T07:16:11.672822+0000 mgr.vm07.yrfcuj (mgr.14201) 1184 : cluster [DBG] pgmap v694: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:13.109 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:12 vm09 bash[21220]: audit 2026-03-07T07:16:11.852844+0000 mgr.vm07.yrfcuj (mgr.14201) 1185 : audit [DBG] from='client.17050 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:13.109 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:12 vm09 bash[21220]: audit 2026-03-07T07:16:11.852844+0000 mgr.vm07.yrfcuj (mgr.14201) 1185 : audit [DBG] from='client.17050 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:13.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:12 vm07 bash[17031]: cluster 2026-03-07T07:16:11.672822+0000 mgr.vm07.yrfcuj (mgr.14201) 1184 : cluster [DBG] pgmap v694: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:13.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:12 vm07 bash[17031]: cluster 2026-03-07T07:16:11.672822+0000 mgr.vm07.yrfcuj (mgr.14201) 1184 : cluster [DBG] pgmap v694: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:13.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:12 vm07 bash[17031]: audit 2026-03-07T07:16:11.852844+0000 mgr.vm07.yrfcuj (mgr.14201) 1185 : audit [DBG] from='client.17050 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:13.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:12 vm07 bash[17031]: audit 2026-03-07T07:16:11.852844+0000 mgr.vm07.yrfcuj (mgr.14201) 1185 : audit [DBG] from='client.17050 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:15.114 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:14 vm09 bash[21220]: cluster 2026-03-07T07:16:13.673257+0000 mgr.vm07.yrfcuj (mgr.14201) 1186 : cluster [DBG] pgmap v695: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:15.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:14 vm09 bash[21220]: cluster 2026-03-07T07:16:13.673257+0000 mgr.vm07.yrfcuj (mgr.14201) 1186 : cluster [DBG] pgmap v695: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:15.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:14 vm07 bash[17031]: cluster 2026-03-07T07:16:13.673257+0000 mgr.vm07.yrfcuj (mgr.14201) 1186 : cluster [DBG] pgmap v695: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:15.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:14 vm07 bash[17031]: cluster 2026-03-07T07:16:13.673257+0000 mgr.vm07.yrfcuj (mgr.14201) 1186 : cluster [DBG] pgmap v695: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:16.694 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T08:16:17.052 INFO:teuthology.orchestra.run.vm07.stdout:HOST ADDR LABELS STATUS 2026-03-07T08:16:17.053 INFO:teuthology.orchestra.run.vm07.stdout:vm07 192.168.123.107 2026-03-07T08:16:17.053 INFO:teuthology.orchestra.run.vm07.stdout:vm09 192.168.123.109 2026-03-07T08:16:17.053 INFO:teuthology.orchestra.run.vm07.stdout:2 hosts in cluster 2026-03-07T08:16:17.064 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:16 vm07 bash[17031]: cluster 2026-03-07T07:16:15.673643+0000 mgr.vm07.yrfcuj (mgr.14201) 1187 : cluster [DBG] pgmap v696: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:17.064 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:16 vm07 bash[17031]: cluster 2026-03-07T07:16:15.673643+0000 mgr.vm07.yrfcuj (mgr.14201) 1187 : cluster [DBG] pgmap v696: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:17.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:16 vm09 bash[21220]: cluster 2026-03-07T07:16:15.673643+0000 mgr.vm07.yrfcuj (mgr.14201) 1187 : cluster [DBG] pgmap v696: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:17.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:16 vm09 bash[21220]: cluster 2026-03-07T07:16:15.673643+0000 mgr.vm07.yrfcuj (mgr.14201) 1187 : cluster [DBG] pgmap v696: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:17.121 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- bash -c 'ceph orch device ls' 2026-03-07T08:16:18.109 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:17 vm09 bash[21220]: audit 2026-03-07T07:16:17.058723+0000 mgr.vm07.yrfcuj (mgr.14201) 1188 : audit [DBG] from='client.17054 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:18.109 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:17 vm09 bash[21220]: audit 2026-03-07T07:16:17.058723+0000 mgr.vm07.yrfcuj (mgr.14201) 1188 : audit [DBG] from='client.17054 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:18.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:17 vm07 bash[17031]: audit 2026-03-07T07:16:17.058723+0000 mgr.vm07.yrfcuj (mgr.14201) 1188 : audit [DBG] from='client.17054 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:18.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:17 vm07 bash[17031]: audit 2026-03-07T07:16:17.058723+0000 mgr.vm07.yrfcuj (mgr.14201) 1188 : audit [DBG] from='client.17054 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:19.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:18 vm09 bash[21220]: cluster 2026-03-07T07:16:17.674089+0000 mgr.vm07.yrfcuj (mgr.14201) 1189 : cluster [DBG] pgmap v697: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:19.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:18 vm09 bash[21220]: cluster 2026-03-07T07:16:17.674089+0000 mgr.vm07.yrfcuj (mgr.14201) 1189 : cluster [DBG] pgmap v697: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:19.145 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:18 vm07 bash[17031]: cluster 2026-03-07T07:16:17.674089+0000 mgr.vm07.yrfcuj (mgr.14201) 1189 : cluster [DBG] pgmap v697: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:19.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:18 vm07 bash[17031]: cluster 2026-03-07T07:16:17.674089+0000 mgr.vm07.yrfcuj (mgr.14201) 1189 : cluster [DBG] pgmap v697: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:21.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:20 vm09 bash[21220]: cluster 2026-03-07T07:16:19.674398+0000 mgr.vm07.yrfcuj (mgr.14201) 1190 : cluster [DBG] pgmap v698: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:21.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:20 vm09 bash[21220]: cluster 2026-03-07T07:16:19.674398+0000 mgr.vm07.yrfcuj (mgr.14201) 1190 : cluster [DBG] pgmap v698: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:21.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:20 vm07 bash[17031]: cluster 2026-03-07T07:16:19.674398+0000 mgr.vm07.yrfcuj (mgr.14201) 1190 : cluster [DBG] pgmap v698: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:21.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:20 vm07 bash[17031]: cluster 2026-03-07T07:16:19.674398+0000 mgr.vm07.yrfcuj (mgr.14201) 1190 : cluster [DBG] pgmap v698: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:21.926 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T08:16:22.284 INFO:teuthology.orchestra.run.vm07.stdout:HOST PATH TYPE DEVICE ID SIZE AVAILABLE REFRESHED REJECT REASONS 2026-03-07T08:16:22.284 INFO:teuthology.orchestra.run.vm07.stdout:vm07 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 18m ago Has a FileSystem, Insufficient space (<5GB) 2026-03-07T08:16:22.284 INFO:teuthology.orchestra.run.vm07.stdout:vm07 /dev/vdb hdd DWNBRSTVMM07001 20.0G No 18m ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T08:16:22.284 INFO:teuthology.orchestra.run.vm07.stdout:vm07 /dev/vdc hdd DWNBRSTVMM07002 20.0G No 18m ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T08:16:22.284 INFO:teuthology.orchestra.run.vm07.stdout:vm07 /dev/vdd hdd DWNBRSTVMM07003 20.0G No 18m ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T08:16:22.284 INFO:teuthology.orchestra.run.vm07.stdout:vm07 /dev/vde hdd DWNBRSTVMM07004 20.0G No 18m ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T08:16:22.284 INFO:teuthology.orchestra.run.vm07.stdout:vm09 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 18m ago Has a FileSystem, Insufficient space (<5GB) 2026-03-07T08:16:22.284 INFO:teuthology.orchestra.run.vm07.stdout:vm09 /dev/vdb hdd DWNBRSTVMM09001 20.0G No 18m ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T08:16:22.284 INFO:teuthology.orchestra.run.vm07.stdout:vm09 /dev/vdc hdd DWNBRSTVMM09002 20.0G No 18m ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T08:16:22.284 INFO:teuthology.orchestra.run.vm07.stdout:vm09 /dev/vdd hdd DWNBRSTVMM09003 20.0G No 18m ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T08:16:22.284 INFO:teuthology.orchestra.run.vm07.stdout:vm09 /dev/vde hdd DWNBRSTVMM09004 20.0G No 18m ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-07T08:16:22.346 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image harbor.clyso.com/custom-ceph/ceph/ceph:cobaltcore-storage-v19.2.3-fasttrack-5 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 -- bash -c 'ceph orch ls | grep '"'"'^osd.all-available-devices '"'"'' 2026-03-07T08:16:23.109 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:22 vm09 bash[21220]: cluster 2026-03-07T07:16:21.674746+0000 mgr.vm07.yrfcuj (mgr.14201) 1191 : cluster [DBG] pgmap v699: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:23.109 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:22 vm09 bash[21220]: cluster 2026-03-07T07:16:21.674746+0000 mgr.vm07.yrfcuj (mgr.14201) 1191 : cluster [DBG] pgmap v699: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:23.109 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:22 vm09 bash[21220]: audit 2026-03-07T07:16:22.290057+0000 mgr.vm07.yrfcuj (mgr.14201) 1192 : audit [DBG] from='client.17058 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:23.109 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:22 vm09 bash[21220]: audit 2026-03-07T07:16:22.290057+0000 mgr.vm07.yrfcuj (mgr.14201) 1192 : audit [DBG] from='client.17058 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:23.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:22 vm07 bash[17031]: cluster 2026-03-07T07:16:21.674746+0000 mgr.vm07.yrfcuj (mgr.14201) 1191 : cluster [DBG] pgmap v699: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:23.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:22 vm07 bash[17031]: cluster 2026-03-07T07:16:21.674746+0000 mgr.vm07.yrfcuj (mgr.14201) 1191 : cluster [DBG] pgmap v699: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:23.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:22 vm07 bash[17031]: audit 2026-03-07T07:16:22.290057+0000 mgr.vm07.yrfcuj (mgr.14201) 1192 : audit [DBG] from='client.17058 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:23.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:22 vm07 bash[17031]: audit 2026-03-07T07:16:22.290057+0000 mgr.vm07.yrfcuj (mgr.14201) 1192 : audit [DBG] from='client.17058 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:25.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:24 vm09 bash[21220]: cluster 2026-03-07T07:16:23.675158+0000 mgr.vm07.yrfcuj (mgr.14201) 1193 : cluster [DBG] pgmap v700: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:25.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:24 vm09 bash[21220]: cluster 2026-03-07T07:16:23.675158+0000 mgr.vm07.yrfcuj (mgr.14201) 1193 : cluster [DBG] pgmap v700: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:25.145 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:24 vm07 bash[17031]: cluster 2026-03-07T07:16:23.675158+0000 mgr.vm07.yrfcuj (mgr.14201) 1193 : cluster [DBG] pgmap v700: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:25.145 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:24 vm07 bash[17031]: cluster 2026-03-07T07:16:23.675158+0000 mgr.vm07.yrfcuj (mgr.14201) 1193 : cluster [DBG] pgmap v700: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 170 B/s rd, 341 B/s wr, 0 op/s 2026-03-07T08:16:26.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:25 vm09 bash[21220]: audit 2026-03-07T07:16:25.149207+0000 mon.vm07 (mon.0) 1338 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:16:26.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:25 vm09 bash[21220]: audit 2026-03-07T07:16:25.149207+0000 mon.vm07 (mon.0) 1338 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:16:26.145 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:25 vm07 bash[17031]: audit 2026-03-07T07:16:25.149207+0000 mon.vm07 (mon.0) 1338 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:16:26.146 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:25 vm07 bash[17031]: audit 2026-03-07T07:16:25.149207+0000 mon.vm07 (mon.0) 1338 : audit [DBG] from='mgr.14201 192.168.123.107:0/1740438509' entity='mgr.vm07.yrfcuj' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-07T08:16:27.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:26 vm09 bash[21220]: cluster 2026-03-07T07:16:25.675568+0000 mgr.vm07.yrfcuj (mgr.14201) 1194 : cluster [DBG] pgmap v701: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:27.115 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:26 vm09 bash[21220]: cluster 2026-03-07T07:16:25.675568+0000 mgr.vm07.yrfcuj (mgr.14201) 1194 : cluster [DBG] pgmap v701: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:27.126 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/mon.vm07/config 2026-03-07T08:16:27.139 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:26 vm07 bash[17031]: cluster 2026-03-07T07:16:25.675568+0000 mgr.vm07.yrfcuj (mgr.14201) 1194 : cluster [DBG] pgmap v701: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:27.139 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:26 vm07 bash[17031]: cluster 2026-03-07T07:16:25.675568+0000 mgr.vm07.yrfcuj (mgr.14201) 1194 : cluster [DBG] pgmap v701: 129 pgs: 129 active+clean; 454 KiB data, 263 MiB used, 160 GiB / 160 GiB avail; 85 B/s rd, 170 B/s wr, 0 op/s 2026-03-07T08:16:27.506 INFO:teuthology.orchestra.run.vm07.stdout:osd.all-available-devices 8 66s ago 21m * 2026-03-07T08:16:27.559 DEBUG:teuthology.run_tasks:Unwinding manager vip 2026-03-07T08:16:27.561 INFO:tasks.vip:Removing 12.12.0.107 (and any VIPs) on vm07.local iface ens3... 2026-03-07T08:16:27.561 DEBUG:teuthology.orchestra.run.vm07:> sudo ip addr del 12.12.0.107/22 dev ens3 2026-03-07T08:16:27.570 DEBUG:teuthology.orchestra.run.vm07:> sudo ip addr del 12.12.1.107/22 dev ens3 2026-03-07T08:16:27.619 INFO:tasks.vip:Removing 12.12.0.109 (and any VIPs) on vm09.local iface ens3... 2026-03-07T08:16:27.619 DEBUG:teuthology.orchestra.run.vm09:> sudo ip addr del 12.12.0.109/22 dev ens3 2026-03-07T08:16:27.626 DEBUG:teuthology.orchestra.run.vm09:> sudo ip addr del 12.12.1.107/22 dev ens3 2026-03-07T08:16:27.676 INFO:teuthology.orchestra.run.vm09.stderr:RTNETLINK answers: Cannot assign requested address 2026-03-07T08:16:27.677 DEBUG:teuthology.orchestra.run:got remote process result: 2 2026-03-07T08:16:27.677 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-07T08:16:27.679 INFO:tasks.cephadm:Teardown begin 2026-03-07T08:16:27.679 DEBUG:teuthology.orchestra.run.vm07:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-07T08:16:27.687 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-07T08:16:27.725 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-07T08:16:27.725 DEBUG:teuthology.orchestra.run.vm07:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-07T08:16:27.730 DEBUG:teuthology.orchestra.run.vm09:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-07T08:16:27.769 INFO:tasks.cephadm:Stopping all daemons... 2026-03-07T08:16:27.769 INFO:tasks.cephadm.mon.vm07:Stopping mon.vm07... 2026-03-07T08:16:27.769 DEBUG:teuthology.orchestra.run.vm07:> sudo systemctl stop ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@mon.vm07 2026-03-07T08:16:27.823 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:27 vm07 systemd[1]: Stopping Ceph mon.vm07 for 312fdbc4-19f2-11f1-81d0-4bbd10a1e012... 2026-03-07T08:16:27.823 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:27 vm07 bash[17031]: audit 2026-03-07T07:16:27.500834+0000 mgr.vm07.yrfcuj (mgr.14201) 1195 : audit [DBG] from='client.17062 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:27.823 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:27 vm07 bash[17031]: audit 2026-03-07T07:16:27.500834+0000 mgr.vm07.yrfcuj (mgr.14201) 1195 : audit [DBG] from='client.17062 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:28.050 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:27 vm07 bash[17031]: debug 2026-03-07T07:16:27.823+0000 7f388b832640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.vm07 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-07T08:16:28.051 INFO:journalctl@ceph.mon.vm07.vm07.stdout:Mar 07 08:16:27 vm07 bash[17031]: debug 2026-03-07T07:16:27.823+0000 7f388b832640 -1 mon.vm07@0(leader) e2 *** Got Signal Terminated *** 2026-03-07T08:16:28.098 DEBUG:teuthology.orchestra.run.vm07:> sudo pkill -f 'journalctl -f -n 0 -u ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@mon.vm07.service' 2026-03-07T08:16:28.109 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:27 vm09 bash[21220]: audit 2026-03-07T07:16:27.500834+0000 mgr.vm07.yrfcuj (mgr.14201) 1195 : audit [DBG] from='client.17062 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:28.110 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 07 08:16:27 vm09 bash[21220]: audit 2026-03-07T07:16:27.500834+0000 mgr.vm07.yrfcuj (mgr.14201) 1195 : audit [DBG] from='client.17062 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-07T08:16:28.124 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-07T08:16:28.124 INFO:tasks.cephadm.mon.vm07:Stopped mon.vm07 2026-03-07T08:16:28.124 INFO:tasks.cephadm.mon.vm09:Stopping mon.vm09... 2026-03-07T08:16:28.124 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@mon.vm09 2026-03-07T08:16:28.290 DEBUG:teuthology.orchestra.run.vm09:> sudo pkill -f 'journalctl -f -n 0 -u ceph-312fdbc4-19f2-11f1-81d0-4bbd10a1e012@mon.vm09.service' 2026-03-07T08:16:28.311 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-07T08:16:28.311 INFO:tasks.cephadm.mon.vm09:Stopped mon.vm09 2026-03-07T08:16:28.311 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 --force --keep-logs 2026-03-07T08:16:28.542 INFO:teuthology.orchestra.run.vm07.stdout:Deleting cluster with fsid: 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T08:17:22.119 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 --force --keep-logs 2026-03-07T08:17:22.362 INFO:teuthology.orchestra.run.vm09.stdout:Deleting cluster with fsid: 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T08:18:14.803 DEBUG:teuthology.orchestra.run.vm07:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-07T08:18:14.810 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-07T08:18:14.817 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-07T08:18:14.817 DEBUG:teuthology.misc:Transferring archived files from vm07:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/crash to /archive/irq0-2026-03-07_00:06:21-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/502/remote/vm07/crash 2026-03-07T08:18:14.818 DEBUG:teuthology.orchestra.run.vm07:> sudo tar c -f - -C /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/crash -- . 2026-03-07T08:18:14.858 INFO:teuthology.orchestra.run.vm07.stderr:tar: /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/crash: Cannot open: No such file or directory 2026-03-07T08:18:14.858 INFO:teuthology.orchestra.run.vm07.stderr:tar: Error is not recoverable: exiting now 2026-03-07T08:18:14.858 DEBUG:teuthology.misc:Transferring archived files from vm09:/var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/crash to /archive/irq0-2026-03-07_00:06:21-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/502/remote/vm09/crash 2026-03-07T08:18:14.858 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/crash -- . 2026-03-07T08:18:14.865 INFO:teuthology.orchestra.run.vm09.stderr:tar: /var/lib/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/crash: Cannot open: No such file or directory 2026-03-07T08:18:14.865 INFO:teuthology.orchestra.run.vm09.stderr:tar: Error is not recoverable: exiting now 2026-03-07T08:18:14.866 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-07T08:18:14.866 DEBUG:teuthology.orchestra.run.vm07:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph.log | egrep CEPHADM_ | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v CEPHADM_DAEMON_PLACE_FAIL | egrep -v CEPHADM_FAILED_DAEMON | head -n 1 2026-03-07T08:18:14.909 INFO:tasks.cephadm:Compressing logs... 2026-03-07T08:18:14.909 DEBUG:teuthology.orchestra.run.vm07:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-07T08:18:14.952 DEBUG:teuthology.orchestra.run.vm09:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-07T08:18:14.959 INFO:teuthology.orchestra.run.vm07.stderr:find: gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-07T08:18:14.959 INFO:teuthology.orchestra.run.vm07.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-07T08:18:14.959 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-client.rgw.foo.vm07.lrxyjy.log 2026-03-07T08:18:14.960 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-client.ceph-exporter.vm07.log 2026-03-07T08:18:14.960 INFO:teuthology.orchestra.run.vm09.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-07T08:18:14.961 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-07T08:18:14.961 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.3.log 2026-03-07T08:18:14.961 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph.log 2026-03-07T08:18:14.963 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.3.log: 90.1% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-07T08:18:14.963 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-client.rgw.foo.vm09.kpgoql.log 2026-03-07T08:18:14.963 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-client.rgw.foo.vm07.lrxyjy.log: 94.0% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-client.rgw.foo.vm07.lrxyjy.log.gz 2026-03-07T08:18:14.963 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph.log 2026-03-07T08:18:14.964 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph.log: 90.4% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph.log.gz 2026-03-07T08:18:14.964 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-client.rgw.foo.vm09.fgzozy.log 2026-03-07T08:18:14.964 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-client.ceph-exporter.vm07.log: 94.8% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-client.ceph-exporter.vm07.log.gz 2026-03-07T08:18:14.965 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-client.rgw.foo.vm07.bgfaka.log 2026-03-07T08:18:14.966 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph.log: 90.3% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph.log.gz 2026-03-07T08:18:14.966 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ops-log-ceph-client.rgw.foo.vm07.bgfaka.log 2026-03-07T08:18:14.966 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-client.rgw.foo.vm09.kpgoql.log: 93.9% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-client.rgw.foo.vm09.kpgoql.log.gz 2026-03-07T08:18:14.967 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.7.log 2026-03-07T08:18:14.969 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-client.rgw.foo.vm07.bgfaka.log: 94.1% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-client.rgw.foo.vm07.bgfaka.log.gz 2026-03-07T08:18:14.969 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-client.rgw.foo.vm09.fgzozy.log: 93.9% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-client.rgw.foo.vm09.fgzozy.log.gz 2026-03-07T08:18:14.969 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.1.log 2026-03-07T08:18:14.969 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ops-log-ceph-client.rgw.foo.vm09.kpgoql.log 2026-03-07T08:18:14.971 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ops-log-ceph-client.rgw.foo.vm07.bgfaka.log: 93.2% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ops-log-ceph-client.rgw.foo.vm07.bgfaka.log.gz 2026-03-07T08:18:14.971 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.5.log 2026-03-07T08:18:14.973 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.1.log: 91.1% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-07T08:18:14.974 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.6.log 2026-03-07T08:18:14.975 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.7.log: gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph.audit.log 2026-03-07T08:18:14.983 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ops-log-ceph-client.rgw.foo.vm09.kpgoql.log: 93.4% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ops-log-ceph-client.rgw.foo.vm09.kpgoql.log.gz 2026-03-07T08:18:14.983 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-volume.log 2026-03-07T08:18:14.985 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph.audit.log: 91.4% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph.audit.log.gz 2026-03-07T08:18:14.992 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.5.log: gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-mgr.vm07.yrfcuj.log 2026-03-07T08:18:14.995 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-mon.vm09.log 2026-03-07T08:18:14.999 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.6.log: gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ops-log-ceph-client.rgw.foo.vm07.lrxyjy.log 2026-03-07T08:18:15.003 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-volume.log: gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ops-log-ceph-client.rgw.foo.vm09.fgzozy.log 2026-03-07T08:18:15.011 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-mon.vm09.log: gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-client.ceph-exporter.vm09.log 2026-03-07T08:18:15.014 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ops-log-ceph-client.rgw.foo.vm09.fgzozy.log: 93.4% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ops-log-ceph-client.rgw.foo.vm09.fgzozy.log.gz 2026-03-07T08:18:15.016 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-mgr.vm07.yrfcuj.log: gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.2.log 2026-03-07T08:18:15.019 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ops-log-ceph-client.rgw.foo.vm07.lrxyjy.log: 93.3%gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph.audit.log 2026-03-07T08:18:15.019 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.2.log: -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ops-log-ceph-client.rgw.foo.vm07.lrxyjy.log.gz 2026-03-07T08:18:15.027 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph.cephadm.log 2026-03-07T08:18:15.028 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-client.ceph-exporter.vm09.log: 31.2% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-client.ceph-exporter.vm09.log.gz 2026-03-07T08:18:15.032 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-volume.log 2026-03-07T08:18:15.038 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph.audit.log: 91.3% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph.audit.log.gz 2026-03-07T08:18:15.043 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.4.log 2026-03-07T08:18:15.043 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph.cephadm.log 2026-03-07T08:18:15.043 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph.cephadm.log: 82.8% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph.cephadm.log.gz 2026-03-07T08:18:15.044 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-mgr.vm09.eqznpw.log 2026-03-07T08:18:15.047 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.4.log: gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.0.log 2026-03-07T08:18:15.050 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-mgr.vm09.eqznpw.log: 93.4% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-volume.log.gz 2026-03-07T08:18:15.052 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-volume.log: gzip -5 --verbose -- /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-mon.vm07.log 2026-03-07T08:18:15.056 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph.cephadm.log: 83.3% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph.cephadm.log.gz 2026-03-07T08:18:15.064 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.0.log: 93.2% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-mgr.vm09.eqznpw.log.gz 2026-03-07T08:18:15.120 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-mon.vm07.log: 93.4% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-volume.log.gz 2026-03-07T08:18:15.383 INFO:teuthology.orchestra.run.vm09.stderr: 92.7% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-mon.vm09.log.gz 2026-03-07T08:18:15.496 INFO:teuthology.orchestra.run.vm07.stderr: 89.8% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-mgr.vm07.yrfcuj.log.gz 2026-03-07T08:18:16.161 INFO:teuthology.orchestra.run.vm07.stderr: 93.4% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.2.log.gz 2026-03-07T08:18:16.204 INFO:teuthology.orchestra.run.vm07.stderr: 90.8% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-mon.vm07.log.gz 2026-03-07T08:18:16.294 INFO:teuthology.orchestra.run.vm09.stderr: 93.4% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.0.log.gz 2026-03-07T08:18:16.294 INFO:teuthology.orchestra.run.vm09.stderr: 93.4% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.7.log.gz 2026-03-07T08:18:16.345 INFO:teuthology.orchestra.run.vm07.stderr: 93.3% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.5.log.gz 2026-03-07T08:18:16.383 INFO:teuthology.orchestra.run.vm07.stderr: 93.3% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.6.log.gz 2026-03-07T08:18:16.398 INFO:teuthology.orchestra.run.vm09.stderr: 93.6% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.4.log.gz 2026-03-07T08:18:16.399 INFO:teuthology.orchestra.run.vm09.stderr: 93.7% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.3.log.gz 2026-03-07T08:18:16.400 INFO:teuthology.orchestra.run.vm09.stderr: 2026-03-07T08:18:16.400 INFO:teuthology.orchestra.run.vm09.stderr:real 0m1.444s 2026-03-07T08:18:16.400 INFO:teuthology.orchestra.run.vm09.stderr:user 0m2.721s 2026-03-07T08:18:16.400 INFO:teuthology.orchestra.run.vm09.stderr:sys 0m0.159s 2026-03-07T08:18:16.594 INFO:teuthology.orchestra.run.vm07.stderr: 93.7% -- replaced with /var/log/ceph/312fdbc4-19f2-11f1-81d0-4bbd10a1e012/ceph-osd.1.log.gz 2026-03-07T08:18:16.596 INFO:teuthology.orchestra.run.vm07.stderr: 2026-03-07T08:18:16.596 INFO:teuthology.orchestra.run.vm07.stderr:real 0m1.642s 2026-03-07T08:18:16.596 INFO:teuthology.orchestra.run.vm07.stderr:user 0m2.894s 2026-03-07T08:18:16.596 INFO:teuthology.orchestra.run.vm07.stderr:sys 0m0.169s 2026-03-07T08:18:16.596 INFO:tasks.cephadm:Archiving logs... 2026-03-07T08:18:16.596 DEBUG:teuthology.misc:Transferring archived files from vm07:/var/log/ceph to /archive/irq0-2026-03-07_00:06:21-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/502/remote/vm07/log 2026-03-07T08:18:16.596 DEBUG:teuthology.orchestra.run.vm07:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-07T08:18:16.805 DEBUG:teuthology.misc:Transferring archived files from vm09:/var/log/ceph to /archive/irq0-2026-03-07_00:06:21-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/502/remote/vm09/log 2026-03-07T08:18:16.805 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-07T08:18:16.959 INFO:tasks.cephadm:Removing cluster... 2026-03-07T08:18:16.959 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 --force 2026-03-07T08:18:17.199 INFO:teuthology.orchestra.run.vm07.stdout:Deleting cluster with fsid: 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T08:18:18.268 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 --force 2026-03-07T08:18:18.512 INFO:teuthology.orchestra.run.vm09.stdout:Deleting cluster with fsid: 312fdbc4-19f2-11f1-81d0-4bbd10a1e012 2026-03-07T08:18:19.577 INFO:tasks.cephadm:Removing cephadm ... 2026-03-07T08:18:19.577 DEBUG:teuthology.orchestra.run.vm07:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-07T08:18:19.580 DEBUG:teuthology.orchestra.run.vm09:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-07T08:18:19.583 INFO:tasks.cephadm:Teardown complete 2026-03-07T08:18:19.583 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-07T08:18:19.585 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-07T08:18:19.586 DEBUG:teuthology.orchestra.run.vm07:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-07T08:18:19.624 DEBUG:teuthology.orchestra.run.vm09:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-07T08:18:20.493 INFO:teuthology.orchestra.run.vm09.stdout: remote refid st t when poll reach delay offset jitter 2026-03-07T08:18:20.493 INFO:teuthology.orchestra.run.vm09.stdout:============================================================================== 2026-03-07T08:18:20.493 INFO:teuthology.orchestra.run.vm09.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T08:18:20.493 INFO:teuthology.orchestra.run.vm09.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T08:18:20.493 INFO:teuthology.orchestra.run.vm09.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T08:18:20.493 INFO:teuthology.orchestra.run.vm09.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T08:18:20.493 INFO:teuthology.orchestra.run.vm09.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T08:18:20.493 INFO:teuthology.orchestra.run.vm09.stdout:*vps-ber1.orlean 127.65.222.189 2 u 75 128 377 28.808 +0.554 0.134 2026-03-07T08:18:20.493 INFO:teuthology.orchestra.run.vm09.stdout:-time.netzwerge. 31.209.85.243 2 u 74 128 377 33.013 +0.127 0.147 2026-03-07T08:18:20.493 INFO:teuthology.orchestra.run.vm09.stdout:-time2.sebhostin 127.65.222.189 2 u 70 128 377 28.813 +0.779 0.219 2026-03-07T08:18:20.493 INFO:teuthology.orchestra.run.vm09.stdout:+node-3.infogral 168.239.11.197 2 u 128 128 377 23.526 +0.432 0.124 2026-03-07T08:18:20.493 INFO:teuthology.orchestra.run.vm09.stdout:+ntp5.kernfusion 237.17.204.95 2 u 72 128 377 28.825 +0.382 0.212 2026-03-07T08:18:20.493 INFO:teuthology.orchestra.run.vm09.stdout:-130.61.89.107 237.17.204.95 2 u 65 128 377 20.958 +0.073 0.117 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout: remote refid st t when poll reach delay offset jitter 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout:============================================================================== 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout:+time.netzwerge. 31.209.85.243 2 u 2 128 377 33.139 -7.267 1.633 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout:+77.90.0.148 (14 131.188.3.220 2 u 63 64 377 22.833 -7.184 2.673 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout:*vps-ber1.orlean 127.65.222.189 2 u - 128 377 28.667 -6.646 1.629 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout:-funky.f5s.de 195.145.119.188 2 u 63 64 377 24.985 -4.713 1.908 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout:-time2.sebhostin 127.65.222.189 2 u 59 128 377 29.002 -4.053 2.025 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout:-time.cloudflare 10.216.8.4 3 u 64 64 377 20.422 -3.604 1.897 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout:#mx03.fischl-onl 185.232.69.65 2 u 3 128 377 24.945 -5.222 2.278 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout:#server1a.meinbe 124.216.164.14 2 u 67 64 377 25.032 -4.706 1.884 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout:#vsrv02141.custo 79.133.44.137 2 u 5 128 377 32.835 -6.465 1.691 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout:#cloudrouter.1in 131.188.3.220 2 u 7 128 377 28.290 -6.827 2.170 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout:-ntp5.kernfusion 237.17.204.95 2 u 3 128 377 28.908 -4.574 2.176 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout:#ntp1.lwlcom.net .GPS. 1 u 1 128 377 30.945 -1.426 2.145 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout:+185.125.190.58 37.15.221.189 2 u 44 128 377 35.843 -5.654 1.977 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout:+130.61.89.107 237.17.204.95 2 u 66 64 377 20.866 -4.657 1.941 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout:#62.108.36.235 ( 40.95.204.66 2 u - 128 377 32.797 -3.433 1.308 2026-03-07T08:18:20.615 INFO:teuthology.orchestra.run.vm07.stdout:+185.125.190.57 194.121.207.249 2 u 39 128 377 34.371 -6.369 2.027 2026-03-07T08:18:20.615 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-07T08:18:20.618 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-07T08:18:20.618 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-07T08:18:20.620 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-07T08:18:20.623 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-07T08:18:20.625 INFO:teuthology.task.internal:Duration was 1684.108429 seconds 2026-03-07T08:18:20.625 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-07T08:18:20.627 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-07T08:18:20.627 DEBUG:teuthology.orchestra.run.vm07:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-07T08:18:20.628 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-07T08:18:20.653 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-07T08:18:20.653 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm07.local 2026-03-07T08:18:20.653 DEBUG:teuthology.orchestra.run.vm07:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-07T08:18:20.705 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm09.local 2026-03-07T08:18:20.705 DEBUG:teuthology.orchestra.run.vm09:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-07T08:18:20.715 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-07T08:18:20.715 DEBUG:teuthology.orchestra.run.vm07:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-07T08:18:20.748 DEBUG:teuthology.orchestra.run.vm09:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-07T08:18:20.820 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-07T08:18:20.821 DEBUG:teuthology.orchestra.run.vm07:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-07T08:18:20.821 DEBUG:teuthology.orchestra.run.vm09:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-07T08:18:20.828 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-07T08:18:20.828 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-07T08:18:20.828 INFO:teuthology.orchestra.run.vm07.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-07T08:18:20.828 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-07T08:18:20.829 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-07T08:18:20.829 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-07T08:18:20.829 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-07T08:18:20.829 INFO:teuthology.orchestra.run.vm09.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-07T08:18:20.829 INFO:teuthology.orchestra.run.vm09.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-07T08:18:20.829 INFO:teuthology.orchestra.run.vm07.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: /home/ubuntu/cephtest/archive/syslog/journalctl.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-07T08:18:20.841 INFO:teuthology.orchestra.run.vm09.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 92.4% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-07T08:18:20.843 INFO:teuthology.orchestra.run.vm07.stderr: 91.4% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-07T08:18:20.844 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-07T08:18:20.847 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-07T08:18:20.847 DEBUG:teuthology.orchestra.run.vm07:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-07T08:18:20.894 DEBUG:teuthology.orchestra.run.vm09:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-07T08:18:20.901 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-07T08:18:20.904 DEBUG:teuthology.orchestra.run.vm07:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-07T08:18:20.936 DEBUG:teuthology.orchestra.run.vm09:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-07T08:18:20.941 INFO:teuthology.orchestra.run.vm07.stdout:kernel.core_pattern = core 2026-03-07T08:18:20.949 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern = core 2026-03-07T08:18:20.957 DEBUG:teuthology.orchestra.run.vm07:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-07T08:18:20.993 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-07T08:18:20.994 DEBUG:teuthology.orchestra.run.vm09:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-07T08:18:21.001 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-07T08:18:21.001 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-07T08:18:21.004 INFO:teuthology.task.internal:Transferring archived files... 2026-03-07T08:18:21.004 DEBUG:teuthology.misc:Transferring archived files from vm07:/home/ubuntu/cephtest/archive to /archive/irq0-2026-03-07_00:06:21-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/502/remote/vm07 2026-03-07T08:18:21.004 DEBUG:teuthology.orchestra.run.vm07:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-07T08:18:21.043 DEBUG:teuthology.misc:Transferring archived files from vm09:/home/ubuntu/cephtest/archive to /archive/irq0-2026-03-07_00:06:21-orch:cephadm:smoke-roleless-cobaltcore-storage-v19.2.3-fasttrack-5-none-default-vps/502/remote/vm09 2026-03-07T08:18:21.043 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-07T08:18:21.050 INFO:teuthology.task.internal:Removing archive directory... 2026-03-07T08:18:21.051 DEBUG:teuthology.orchestra.run.vm07:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-07T08:18:21.084 DEBUG:teuthology.orchestra.run.vm09:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-07T08:18:21.098 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-07T08:18:21.100 INFO:teuthology.task.internal:Not uploading archives. 2026-03-07T08:18:21.100 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-07T08:18:21.103 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-07T08:18:21.103 DEBUG:teuthology.orchestra.run.vm07:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-07T08:18:21.128 DEBUG:teuthology.orchestra.run.vm09:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-07T08:18:21.130 INFO:teuthology.orchestra.run.vm07.stdout: 258079 4 drwxr-xr-x 2 ubuntu ubuntu 4096 Mar 7 08:18 /home/ubuntu/cephtest 2026-03-07T08:18:21.142 INFO:teuthology.orchestra.run.vm09.stdout: 258067 4 drwxr-xr-x 2 ubuntu ubuntu 4096 Mar 7 08:18 /home/ubuntu/cephtest 2026-03-07T08:18:21.142 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-07T08:18:21.149 INFO:teuthology.run:Summary data: description: orch:cephadm:smoke-roleless/{0-distro/ubuntu_22.04 1-start 2-services/rgw-ingress 3-final} duration: 1684.1084289550781 owner: irq0 success: true 2026-03-07T08:18:21.149 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-07T08:18:21.166 INFO:teuthology.run:pass