squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/osds/{0-distro/centos_9.stream_runc 0-nvme-loop 1-start 2-ops/rm-zap-flag}
"2026-03-08T22:47:11.506052+0000 mon.vm08 (mon.0) 494 : cluster [WRN] Health check failed: Failed to apply 1 service(s): osd.all-available-devices (CEPHADM_APPLY_SPEC_FAIL)" in cluster log
squid
tt-squid
clyso-debian-13
vps
ubuntu 22.04
orch:cephadm/smoke-roleless/{0-distro/ubuntu_22.04 0-nvme-loop 1-start 2-services/nfs-ingress-rgw-user 3-final}
Command failed on vm02 with status 1: "grep '^nvme_loop' /proc/modules || sudo modprobe nvme_loop && sudo mkdir -p /sys/kernel/config/nvmet/hosts/hostnqn && sudo mkdir -p /sys/kernel/config/nvmet/ports/1 && echo loop | sudo tee /sys/kernel/config/nvmet/ports/1/addr_trtype"
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/smoke/{0-distro/centos_9.stream_runc 0-nvme-loop agent/off fixed-2 mon_election/classic start}
squid
tt-squid
clyso-debian-13
vps
ubuntu 22.04
orch:cephadm/with-work/{0-distro/ubuntu_22.04 fixed-2 mode/root mon_election/connectivity msgr/async-v2only start tasks/rotate-keys}
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_9.stream conf/{client mds mgr mon osd} fail_fs/no kernel overrides/{ignorelist_health ignorelist_upgrade ignorelist_wrongly_marked_down pg-warn pg_health syntax} roles tasks/{0-from/quincy 1-volume/{0-create 1-ranks/2 2-allow_standby_replay/yes 3-inline/no 4-verify} 2-client/fuse 3-upgrade-mgr-staggered 4-config-upgrade/{fail_fs} 5-upgrade-with-workload 6-verify}}
"grep: /var/log/ceph/178106ac-1b42-11f1-8095-5f48055c15ba/ceph.log: No such file or directory" in cluster log
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/mgr-nfs-upgrade/{0-centos_9.stream 1-bootstrap/17.2.0 1-start 2-nfs 3-upgrade-with-workload 4-final}
squid
tt-squid
clyso-debian-13
vps
ubuntu 22.04
orch:cephadm/nfs/{cluster/{1-node} conf/{client mds mgr mon osd} overrides/{ignore_mgr_down ignorelist_health pg_health} supported-random-distros$/{ubuntu_latest} tasks/nfs}
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/no-agent-workunits/{0-distro/centos_9.stream mon_election/classic task/test_orch_cli}
squid
tt-squid
clyso-debian-13
vps
ubuntu 22.04
orch:cephadm/orchestrator_cli/{0-random-distro$/{ubuntu_22.04} 2-node-mgr agent/off orchestrator_cli}
squid
tt-squid
clyso-debian-13
vps
ubuntu 22.04
orch:cephadm/rbd_iscsi/{base/install cluster/{fixed-3 openstack} conf/{disable-pool-app} supported-container-hosts$/{ubuntu_22.04} workloads/cephadm_iscsi}
Command failed on vm02 with status 1: 'CEPH_REF=master CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/virtualenv/bin/cram -v -- /home/ubuntu/cephtest/archive/cram.client.0/*.t'
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/smb/{0-distro/centos_9.stream tasks/deploy_smb_basic}
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/smoke-roleless/{0-distro/centos_9.stream 0-nvme-loop 1-start 2-services/nfs-ingress 3-final}
"2026-03-08T23:21:03.742359+0000 mon.vm06 (mon.0) 492 : cluster [WRN] Health check failed: Failed to apply 1 service(s): osd.all-available-devices (CEPHADM_APPLY_SPEC_FAIL)" in cluster log
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/smoke-singlehost/{0-random-distro$/{centos_9.stream} 1-start 2-services/basic 3-final}
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/smoke-small/{0-distro/centos_9.stream_runc 0-nvme-loop agent/off fixed-2 mon_election/classic start}
squid
tt-squid
clyso-debian-13
vps
ubuntu 22.04
orch:cephadm/workunits/{0-distro/ubuntu_22.04 agent/off mon_election/classic task/test_extra_daemon_features}
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/smoke-roleless/{0-distro/centos_9.stream_runc 0-nvme-loop 1-start 2-services/nfs-ingress2 3-final}
"2026-03-08T23:46:25.629198+0000 mon.vm04 (mon.0) 496 : cluster [WRN] Health check failed: Failed to apply 1 service(s): osd.all-available-devices (CEPHADM_APPLY_SPEC_FAIL)" in cluster log
squid
tt-squid
clyso-debian-13
vps
ubuntu 22.04
orch:cephadm/osds/{0-distro/ubuntu_22.04 0-nvme-loop 1-start 2-ops/rm-zap-wait}
Command failed on vm01 with status 1: "grep '^nvme_loop' /proc/modules || sudo modprobe nvme_loop && sudo mkdir -p /sys/kernel/config/nvmet/hosts/hostnqn && sudo mkdir -p /sys/kernel/config/nvmet/ports/1 && echo loop | sudo tee /sys/kernel/config/nvmet/ports/1/addr_trtype"
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/thrash/{0-distro/centos_9.stream 1-start 2-thrash 3-tasks/small-objects fixed-2 msgr/async-v1only root}
squid
tt-squid
clyso-debian-13
vps
ubuntu 22.04
orch:cephadm/smoke-roleless/{0-distro/ubuntu_22.04 0-nvme-loop 1-start 2-services/nfs-keepalive-only 3-final}
Command failed on vm01 with status 1: "grep '^nvme_loop' /proc/modules || sudo modprobe nvme_loop && sudo mkdir -p /sys/kernel/config/nvmet/hosts/hostnqn && sudo mkdir -p /sys/kernel/config/nvmet/ports/1 && echo loop | sudo tee /sys/kernel/config/nvmet/ports/1/addr_trtype"
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/workunits/{0-distro/centos_9.stream agent/on mon_election/connectivity task/test_host_drain}
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/smoke-roleless/{0-distro/centos_9.stream 0-nvme-loop 1-start 2-services/nfs 3-final}
"2026-03-08T23:50:37.353868+0000 mon.vm01 (mon.0) 500 : cluster [WRN] Health check failed: Failed to apply 1 service(s): osd.all-available-devices (CEPHADM_APPLY_SPEC_FAIL)" in cluster log
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/upgrade/{1-start-distro/1-start-centos_9.stream 2-repo_digest/defaut 3-upgrade/staggered 4-wait 5-upgrade-ls agent/off mon_election/classic}
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/mds_upgrade_sequence/{bluestore-bitmap centos_9.stream conf/{client mds mgr mon osd} fail_fs/yes kernel overrides/{ignorelist_health ignorelist_upgrade ignorelist_wrongly_marked_down pg-warn pg_health syntax} roles tasks/{0-from/reef/{v18.2.1} 1-volume/{0-create 1-ranks/1 2-allow_standby_replay/no 3-inline/yes 4-verify} 2-client/kclient 3-upgrade-mgr-staggered 4-config-upgrade/{fail_fs} 5-upgrade-with-workload 6-verify}}
"2026-03-09T00:10:00.000194+0000 mon.vm03 (mon.0) 518 : cluster [WRN] osd.3 (root=default,host=vm06) is down" in cluster log
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/no-agent-workunits/{0-distro/centos_9.stream_runc mon_election/connectivity task/test_orch_cli_mon}
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/smb/{0-distro/centos_9.stream_runc tasks/deploy_smb_domain}
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/smoke-roleless/{0-distro/centos_9.stream_runc 0-nvme-loop 1-start 2-services/nfs2 3-final}
"2026-03-09T00:28:04.785200+0000 mon.vm02 (mon.0) 495 : cluster [WRN] Health check failed: Failed to apply 1 service(s): osd.all-available-devices (CEPHADM_APPLY_SPEC_FAIL)" in cluster log
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/with-work/{0-distro/centos_9.stream fixed-2 mode/packaged mon_election/classic msgr/async-v2only start tasks/rados_api_tests}
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/workunits/{0-distro/centos_9.stream_runc agent/off mon_election/classic task/test_iscsi_container/{centos_9.stream test_iscsi_container}}
squid
tt-squid
clyso-debian-13
vps
centos 9.stream
orch:cephadm/osds/{0-distro/centos_9.stream 0-nvme-loop 1-start 2-ops/rmdir-reactivate}
squid
tt-squid
clyso-debian-13
vps
ubuntu 22.04
orch:cephadm/smoke-roleless/{0-distro/ubuntu_22.04 0-nvme-loop 1-start 2-services/nvmeof 3-final}
Command failed on vm03 with status 1: "grep '^nvme_loop' /proc/modules || sudo modprobe nvme_loop && sudo mkdir -p /sys/kernel/config/nvmet/hosts/hostnqn && sudo mkdir -p /sys/kernel/config/nvmet/ports/1 && echo loop | sudo tee /sys/kernel/config/nvmet/ports/1/addr_trtype"
squid
tt-squid
clyso-debian-13
vps
ubuntu 22.04
orch:cephadm/smoke/{0-distro/ubuntu_22.04 0-nvme-loop agent/on fixed-2 mon_election/connectivity start}