ID
Status
Ceph Branch
Suite Branch
Teuthology Branch
Machine
OS
Nodes
Description
Failure Reason
cobaltcore-storage-v19.2.3-fasttrack-3
tt-19.2.3-fasttrack-3
clyso-debian-13
vps
centos 9.stream
orch:cephadm:osds/{0-distro/centos_9.stream 0-nvme-loop 1-start 2-ops/rmdir-reactivate}
"2026-03-05T22:47:38.385867+0000 mon.vm02 (mon.0) 504 : cluster [WRN] Health check failed: Failed to apply 1 service(s): osd.all-available-devices (CEPHADM_APPLY_SPEC_FAIL)" in cluster log
cobaltcore-storage-v19.2.3-fasttrack-3
tt-19.2.3-fasttrack-3
clyso-debian-13
vps
centos 9.stream
orch:cephadm:osds/{0-distro/centos_9.stream_runc 0-nvme-loop 1-start 2-ops/repave-all}
"2026-03-05T22:47:43.365766+0000 mon.vm01 (mon.0) 500 : cluster [WRN] Health check failed: Failed to apply 1 service(s): osd.all-available-devices (CEPHADM_APPLY_SPEC_FAIL)" in cluster log
cobaltcore-storage-v19.2.3-fasttrack-3
tt-19.2.3-fasttrack-3
clyso-debian-13
vps
ubuntu 22.04
orch:cephadm:osds/{0-distro/ubuntu_22.04 0-nvme-loop 1-start 2-ops/rm-zap-add}
Command failed on vm06 with status 1: "grep '^nvme_loop' /proc/modules || sudo modprobe nvme_loop && sudo mkdir -p /sys/kernel/config/nvmet/hosts/hostnqn && sudo mkdir -p /sys/kernel/config/nvmet/ports/1 && echo loop | sudo tee /sys/kernel/config/nvmet/ports/1/addr_trtype"
cobaltcore-storage-v19.2.3-fasttrack-3
tt-19.2.3-fasttrack-3
clyso-debian-13
vps
centos 9.stream
orch:cephadm:osds/{0-distro/centos_9.stream 0-nvme-loop 1-start 2-ops/rm-zap-flag}
"2026-03-05T22:49:30.995739+0000 mon.vm00 (mon.0) 502 : cluster [WRN] Health check failed: Failed to apply 1 service(s): osd.all-available-devices (CEPHADM_APPLY_SPEC_FAIL)" in cluster log
cobaltcore-storage-v19.2.3-fasttrack-3
tt-19.2.3-fasttrack-3
clyso-debian-13
vps
centos 9.stream
orch:cephadm:osds/{0-distro/centos_9.stream_runc 0-nvme-loop 1-start 2-ops/rm-zap-wait}
"2026-03-05T22:52:16.375038+0000 mon.vm06 (mon.0) 497 : cluster [WRN] Health check failed: Failed to apply 1 service(s): osd.all-available-devices (CEPHADM_APPLY_SPEC_FAIL)" in cluster log
cobaltcore-storage-v19.2.3-fasttrack-3
tt-19.2.3-fasttrack-3
clyso-debian-13
vps
ubuntu 22.04
orch:cephadm:osds/{0-distro/ubuntu_22.04 0-nvme-loop 1-start 2-ops/rmdir-reactivate}
Command failed on vm03 with status 1: "grep '^nvme_loop' /proc/modules || sudo modprobe nvme_loop && sudo mkdir -p /sys/kernel/config/nvmet/hosts/hostnqn && sudo mkdir -p /sys/kernel/config/nvmet/ports/1 && echo loop | sudo tee /sys/kernel/config/nvmet/ports/1/addr_trtype"
cobaltcore-storage-v19.2.3-fasttrack-3
tt-19.2.3-fasttrack-3
clyso-debian-13
vps
centos 9.stream
orch:cephadm:osds/{0-distro/centos_9.stream 0-nvme-loop 1-start 2-ops/repave-all}
"2026-03-05T22:24:15.620263+0000 mon.vm06 (mon.0) 492 : cluster [WRN] Health check failed: Failed to apply 1 service(s): osd.all-available-devices (CEPHADM_APPLY_SPEC_FAIL)" in cluster log
cobaltcore-storage-v19.2.3-fasttrack-3
tt-19.2.3-fasttrack-3
clyso-debian-13
vps
centos 9.stream
orch:cephadm:osds/{0-distro/centos_9.stream_runc 0-nvme-loop 1-start 2-ops/rm-zap-add}
"2026-03-05T22:25:02.513132+0000 mon.vm00 (mon.0) 497 : cluster [WRN] Health check failed: Failed to apply 1 service(s): osd.all-available-devices (CEPHADM_APPLY_SPEC_FAIL)" in cluster log
cobaltcore-storage-v19.2.3-fasttrack-3
tt-19.2.3-fasttrack-3
clyso-debian-13
vps
ubuntu 22.04
orch:cephadm:osds/{0-distro/ubuntu_22.04 0-nvme-loop 1-start 2-ops/rm-zap-flag}
Command failed on vm02 with status 1: "grep '^nvme_loop' /proc/modules || sudo modprobe nvme_loop && sudo mkdir -p /sys/kernel/config/nvmet/hosts/hostnqn && sudo mkdir -p /sys/kernel/config/nvmet/ports/1 && echo loop | sudo tee /sys/kernel/config/nvmet/ports/1/addr_trtype"
cobaltcore-storage-v19.2.3-fasttrack-3
tt-19.2.3-fasttrack-3
clyso-debian-13
vps
centos 9.stream
orch:cephadm:osds/{0-distro/centos_9.stream 0-nvme-loop 1-start 2-ops/rm-zap-wait}
"2026-03-05T22:28:25.815223+0000 mon.vm03 (mon.0) 499 : cluster [WRN] Health check failed: Failed to apply 1 service(s): osd.all-available-devices (CEPHADM_APPLY_SPEC_FAIL)" in cluster log
cobaltcore-storage-v19.2.3-fasttrack-3
tt-19.2.3-fasttrack-3
clyso-debian-13
vps
centos 9.stream
orch:cephadm:osds/{0-distro/centos_9.stream_runc 0-nvme-loop 1-start 2-ops/rmdir-reactivate}
"2026-03-05T22:32:24.623279+0000 mon.vm02 (mon.0) 496 : cluster [WRN] Health check failed: Failed to apply 1 service(s): osd.all-available-devices (CEPHADM_APPLY_SPEC_FAIL)" in cluster log
cobaltcore-storage-v19.2.3-fasttrack-3
tt-19.2.3-fasttrack-3
clyso-debian-13
vps
ubuntu 22.04
orch:cephadm:osds/{0-distro/ubuntu_22.04 0-nvme-loop 1-start 2-ops/repave-all}
Command failed on vm06 with status 1: "grep '^nvme_loop' /proc/modules || sudo modprobe nvme_loop && sudo mkdir -p /sys/kernel/config/nvmet/hosts/hostnqn && sudo mkdir -p /sys/kernel/config/nvmet/ports/1 && echo loop | sudo tee /sys/kernel/config/nvmet/ports/1/addr_trtype"
cobaltcore-storage-v19.2.3-fasttrack-3
tt-19.2.3-fasttrack-3
clyso-debian-13
vps
centos 9.stream
orch:cephadm:osds/{0-distro/centos_9.stream 0-nvme-loop 1-start 2-ops/rm-zap-add}
"2026-03-05T22:36:23.612111+0000 mon.vm00 (mon.0) 499 : cluster [WRN] Health check failed: Failed to apply 1 service(s): osd.all-available-devices (CEPHADM_APPLY_SPEC_FAIL)" in cluster log
cobaltcore-storage-v19.2.3-fasttrack-3
tt-19.2.3-fasttrack-3
clyso-debian-13
vps
centos 9.stream
orch:cephadm:osds/{0-distro/centos_9.stream_runc 0-nvme-loop 1-start 2-ops/rm-zap-flag}
"2026-03-05T22:41:45.425799+0000 mon.vm03 (mon.0) 501 : cluster [WRN] Health check failed: Failed to apply 1 service(s): osd.all-available-devices (CEPHADM_APPLY_SPEC_FAIL)" in cluster log
cobaltcore-storage-v19.2.3-fasttrack-3
tt-19.2.3-fasttrack-3
clyso-debian-13
vps
ubuntu 22.04
orch:cephadm:osds/{0-distro/ubuntu_22.04 0-nvme-loop 1-start 2-ops/rm-zap-wait}
Command failed on vm06 with status 1: "grep '^nvme_loop' /proc/modules || sudo modprobe nvme_loop && sudo mkdir -p /sys/kernel/config/nvmet/hosts/hostnqn && sudo mkdir -p /sys/kernel/config/nvmet/ports/1 && echo loop | sudo tee /sys/kernel/config/nvmet/ports/1/addr_trtype"